Denys88/rl_games

Why is the performance of GRU+PPO poor?

Closed this issue · 4 comments

The performance of your project is shocking. I want to know why CNN+PPO can be so excellent.
Thank you!!!!!!

hi @hijkzzz the most weird part that tensorflow performs much better on MMM2 with conv1d than pytorch.
If you take a look at DM/sc2_runs you can see that tenosrflow:
'python runner.py --train --file rl_games/configs/smac/MMM2.yaml --tf' trains in a few millions of runs but
'python runner.py --train --file rl_games/configs/smac/MMM2_torch.yaml' trains much slover. But I even reproduced truncated normal initializer and still have the same issues.

But I've found that simple MLP is enough to solve most of the envs.
You can find configs in 'rl_games/configs/smac/runs'.
I'll merge DM/sc2_runs in afew days. If you got any issues in current branch they gonna be fixed here.

hi @hijkzzz the most weird part that tensorflow performs much better on MMM2 with conv1d than pytorch.
If you take a look at DM/sc2_runs you can see that tenosrflow:
'python runner.py --train --file rl_games/configs/smac/MMM2.yaml --tf' trains in a few millions of runs but
'python runner.py --train --file rl_games/configs/smac/MMM2_torch.yaml' trains much slover. But I even reproduced truncated normal initializer and still have the same issues.

But I've found that simple MLP is enough to solve most of the envs.
You can find configs in 'rl_games/configs/smac/runs'.
I'll merge DM/sc2_runs in afew days. If you got any issues in current branch they gonna be fixed here.

Thank you, I test RNN+PPO in SMAC, but get poor performance, so I am particularly curious about which tricks make MLP or CNN+ppo work.

@hijkzzz I've found a few bugs some time ago and fixed them. Going to rerun sum rnn experiments.

@hijkzzz I've found a few more bugs in the RNN implementation, now it works much better.