Implementation tricks for different algorithms
waterhorse1 opened this issue · 2 comments
It's a really Nice work on benchmarking multi-agent RL algorithms and I really like it. When I go through the code, I find out epymarl basically only implements the basic version of different algorithms and ignore many different implementation tricks for different algorithms, such as the value normalization trick for MAPPO mentioned in MAPPO , or the value clipping trick for IPPO mentioned in IPPO. Will furture version epymarl support such tricks?
Hello,
we aim to add more implementation details of different algorithms in the future. The basic idea was for all algorithms to have very similar implementations to be able to better compare their achieved returns. We are also happy to accept pull requests for these implementation tricks.
Thank you for your answer.