Issues
- 0
- 1
- 2
Chess and other non-trivial games
#231 opened by StepHaze - 1
Sampled MuZero implementation
#191 opened by matthiaskiller - 2
Only One Player: Can we use MuZero?
#210 opened by 1121091694 - 2
Without Selfplay, why 1 games on a running??
#217 opened by dlrlfkr11 - 1
What does the `replay_buffer.pkl` do?
#228 opened by dmtrung14 - 0
cross play between models
#229 opened by dmtrung14 - 11
The model does not converge for breakout
#211 opened by yungangwu - 1
How can I use a pre-trained model?
#227 opened by worldsoft - 1
Update gym package to gymnasium
#219 opened by Mlokos - 0
MuZero choose the same action
#226 opened by sdumi03 - 0
Muzero crashes when choosing spiel/backgammon
#225 opened by artshar - 0
question about action encoded
#222 opened by Nightbringers - 0
The difference between offical pseudo code and this repository about "num_unroll_steps"
#221 opened by ZF4444 - 0
Dirichlet noise added outside of training
#220 opened by TommyX12 - 0
Mean_value plot in Total_reward - Interpretation
#215 opened by SunilaAkbar - 0
- 2
Can't train using GPU? The torch version for this environment is '1.10.0cpu', that is, CPU one.
#213 opened by SunilaAkbar - 0
Question about the perspective transformation of two players when calculating Q?
#212 opened by puyuan1996 - 0
TypeError: can't pickle function objects
#208 opened by OopsYouDiedE - 7
MuZero Unplugged
#185 opened by tbskrpmnns - 3
Question: Does muzero-general support 2 player games with simultaneous action selection?
#207 opened by moscoso - 1
ray.exceptions.RayActorError: The actor died unexpectedly before finishing this task.
#182 opened by hairinwind - 1
Render model
#197 opened by theeduardomora - 0
Remove Batch Norm?
#203 opened by verbose-void - 0
OpenGL rendering on a remote server over X11
#202 opened by jrjbertram - 3
RuntimeError: module must have its parameters and buffers on device cuda:0 (device_ids[0]) but found one of them on device: cpu
#174 opened by lukaszkn - 0
- 1
raw install has ray problem
#198 opened by EngrStudent - 0
sampling in continuous/complex action spaces with 'density prior' is not working
#200 opened by ManorZ - 0
Batch MCTS
#199 opened by szrlee - 1
- 0
custom observation transformation
#195 opened by SimpleMathmatics - 1
- 1
Dimensionality issue in continuous action space
#165 opened by alik604 - 1
Target Value Offset
#170 opened by dans-acc - 1
Scaling of historical stacked observations
#178 opened by tuero - 1
training result cannot be loaded on another machine
#180 opened by hairinwind - 4
Entropy loss in continuous actions
#175 opened by 2M-kotb - 1
- 0
Struggling to get Ray working
#171 opened by SheldonCurtiss - 0
Why is root.visit_count initialized to 0 and root_predicted_value not included in root node value?
#184 opened by dniku - 0
- 3
Why my game does not remember the steps trained
#181 opened by hairinwind - 2
Total Training Reward rises then drops again
#176 opened by annahambi - 1
Is there a more optimized way to complete training? Every game has to be learned for so long, people learn to master it early
#173 opened by lwaif - 2
- 1
- 2
could it run without ray?
#164 opened by hilberthu