Issues
- 1
- 1
- 1
- 0
MADDPG for custom environment
#82 opened by kapilgarg7568 - 2
Loading previous state...
#70 opened by yanwaiwai - 0
python3.5.4已经进行维护。
#80 opened by vducd - 0
- 6
TypeError: set_color() got multiple values for argument 'alpha' in Simple-Crypto
#30 opened by marwanihab - 1
what's benchmark used for?
#35 opened by KK666-AI - 0
- 2
- 1
- 5
Two problem about update function
#32 opened by YuanyeMa - 0
- 1
Question about the way how to update actor
#65 opened by choasLC - 0
A question about maddpg.py
#68 opened by chengdusunny - 3
There is no provision to run ddpg.
#33 opened by frenzytejask98 - 1
Training EVERY step, not every 100
#67 opened by eflopez1 - 1
SoftMultiCategoricalPd
#40 opened by sandeepnRES - 1
How to normalize the data in table of Appendix to obtain Figure 3 in paper?
#55 opened by flammingRaven - 0
how to evaluate maddpg?
#57 opened by Xinlei-Ren - 1
The code does not converged
#34 opened by sjq19960802 - 0
How to run ddpg
#64 opened by EvaluationResearch - 1
- 0
- 0
../../..
#59 opened by devarani10 - 0
Using any scenario rather than the "simple" one gives error during loading the model after training
#58 opened by dr-smgad - 2
- 2
Can this algorithm be generalised to work with multiple (60) agents competing against eachother?
#27 opened by alexanderkell - 0
how to evaluate maddpg?
#56 opened by Xinlei-Ren - 3
Nontype flaw in "train.py", line 182
#28 opened by DailinH - 0
How can I use DDPG to train it?
#53 opened by CHH3213 - 0
TypeError: must be str, not NoneType
#52 opened by CHH3213 - 0
reward is too large
#51 opened by Sherry-97 - 2
- 0
The sample function in distribution is implementation of Gumbel-softmax, I added it to my code, now it helps to speed up stabilize the training, but my speaker still can not tell the different landmarks.
#48 opened by tanxiangtj - 0
Episode in cooperative navigation env
#47 opened by kargarisaac - 0
question about p_reg in p_train
#46 opened by yeshenpy - 1
Hello! I encountered some problems while running the train.py file under the MADDPG file and would like to seek your help.
#36 opened by dcy0324 - 0
Typo in train.py
#44 opened by opt12 - 0
Question regarding the replay buffers and the Critic networks. (duplicates in the state)
#43 opened by opt12 - 1
run code
#39 opened by lionel-xie - 0
Spark
#41 opened by diemanalytics-ewd - 3
Error when setting display to true
#31 opened by njfdiem - 3
- 1
Having trouble with import maddpg
#22 opened by ishanivyas - 2
- 0
Cumulative rewards are not promoted when use MADDPG
#26 opened by jhcknzzm - 0
Please add a description to this repo
#21 opened by clintonyeb - 1
Calculating Success Rate for Physical Deception
#20 opened by ZishunYu