Issues
- 1
Is this code SAC-V not a SAC?
#43 opened by night2570 - 2
[Question] Mask Batch
#42 opened by chenxi-yang - 0
Question: Why optimize loss_alpha?
#47 opened by DefinitlyEvil - 0
- 0
No normalization of state space
#46 opened by rosa-wolf - 0
- 1
Model saving and loading
#41 opened by tissten - 1
- 3
- 3
- 0
question about q_loss and alpha_loss
#40 opened by xxxkxin - 0
Doubts about Regularization in policy loss
#39 opened by Marxvans - 5
Resume training
#35 opened by Tomeu7 - 6
- 10
Exploding entropy temperature
#34 opened by reubenwong97 - 1
- 1
- 2
Inconsistent seeding
#32 opened by mohakbhardwaj - 1
Action scale and action bias
#24 opened by shakenov-chinga - 0
Support OpenAI Gym Robotic Env?
#30 opened by peiseng - 4
Target value calculation maistake
#25 opened by alirezakazemipour - 4
- 1
Can I use this in custom gym env?
#26 opened by kwk2696 - 4
Policy Loss with Minimum or Q1?
#3 opened by pranv - 4
- 2
puzzles about action scaling
#20 opened by wayunderfoot - 11
Question about policy_loss
#14 opened by toshikwa - 3
About model.py line 105
#17 opened by BangLiu - 9
Derivative in reparametrization trick?
#11 opened by ZeratuuLL - 4
Value network
#9 opened by jendelel - 3
Normalized Actions has bugs
#12 opened by Phlogiston90 - 5
reproducibility for HalfCheetah-v2
#4 opened by tldoan - 1
- 4
Why do you need to use NormalizedActions()?
#6 opened by JingJerry - 10
reparametrization trick issue
#5 opened by tldoan - 3
A question in the deterministic case
#2 opened by roosephu - 1