/EMU

(Official) PyTorch implementation for Efficient Episodic Memory Utilization of Cooperative Multi-Agent Reinforcement Learning (EMU) in ICLR 2024.

Primary LanguagePythonApache License 2.0Apache-2.0

EMU: Efficient Episodic Memory Utilization of Cooperative Multi-agent Reinforcement Learning

Note

This codebase accompanies the paper submission "Efficient Episodic Memory Utilization of Cooperative Multi-agent Reinforcement Learning (EMU)" and is based on GRF, PyMARL and SMAC which are open-sourced. The paper is accepted by ICLR2024 and now available in OpenReview and arXiv.

PyMARL is WhiRL's framework for deep multi-agent reinforcement learning and our code includes implementations of the following algorithms:

Run an experiment

To train EMU(QPLEX) on SC2 setting tasks, run the following command:

python3 src/main.py --config=EMU_sc2 --env-config=sc2 with env_args.map_name=5m_vs_6m

For EMU(CDS), please change config file to EMU_sc2_cds.

To train EMU(QPLEX) on SC2 setting tasks, run the following command:

python3 src/main.py --config=EMU_grf --env-config=academy_3_vs_1_with_keeper

For EMU(CDS), please change config file to EMU_grf_cds. (Note: please set optimality_incentive=True for episodic incentive in EMU_CDS configurations. We will update the default configuration soon.)

Publication

If you find this repository useful, please cite our paper:

@inproceedings{na2024efficient,
  title={Efficient Episodic Memory Utilization of Cooperative Multi-Agent Reinforcement Learning},
  author={Na, Hyungho and Seo, Yunkyeong and Moon, Il-chul},
  booktitle={The Twelfth International Conference on Learning Representations},
  year={2024}
}