/MARL

Implementation for mSAC methods in PyTorch

Primary LanguagePython

StarCraft II micromanagement

Our code is modified from https://github.com/starry-sky6688/StarCraft.

This repository is our Pytorch implementations for Decomposed Soft Actor-Critic Method for Cooperative Multi-Agent Reinforcement Learning(https://arxiv.org/abs/2104.06655).

In the paper, we incorporates the idea of the multi-agent value function decomposition and soft actor-critic framework effectively and proposes a new method mSAC. Experimental results demonstrate that mSAC significantly outperforms policy-based approach–COMA, and achieves competitive performance with SOTA value-based approach–Qmix on most tasks in terms of asymptotic performance metric. In addition, mSAC has achieved significantly better results than Qmix in some tasks with large action spaces (such as 2c_vs_64zg, MMM2).

We trained these algorithms on SMAC, the decentralised micromanagement scenario of StarCraft II.

Corresponding Papers

Requirements

Acknowledgement

Quick Start

$ python src/main.py --map=3m

To run different variant algorithms, you can change the first line in src/main.py [from runner_msac import Runner] to different runner_[alg. name].