RACE icon indicating copy to clipboard operation
RACE copied to clipboard

(ICML 2023) The official code for RACE: Improve Multi-Agent Reinforcement Learning with Representation Asymmetry and Collaborative Evolution

RACE: Improve Multi-Agent Reinforcement Learning with Representation Asymmetry and Collaborative Evolution (ICML 2023)

This repo contains the code that was used in the paper "RACE: Improve Multi-Agent Reinforcement Learning with Representation Asymmetry and Collaborative Evolution". It includes implementations for RACE, FACMAC, FACMAC-nonmonotonic, MADDPG, COMIX, COVDN, and Independent DDPG. Our code is built upon the FACMAC codebase.

RACE code for MA-MuJoCo is released: https://github.com/yeshenpy/RACE-MAMuJoCo

Although the PYMARL framework can be used to run SMAC as well as MAMUJOCO, we do not run MAMUJOCO in this framework.

If you are interested in Combing Evolutionary Algorithms and Reinforcement Learning, you can access this repository Awesome-Evolutionary-Reinforcement-Learning for advanced ERL works.





[Paper]

Setup instructions

Set up StarCraft II and SMAC:

bash install_sc2.sh

Install the dependent packages

pip install -r requirements.txt

Our code uses WandB for visualization. Before you run it, please configure WandB.

Environments

StarCraft Multi-Agent Challenge (SMAC)

We use the SMAC environment developed by WhiRL. Please check the SMAC repo for more details about the environment. Note that for all SMAC experiments, we used SC2.4.10.

Run an experiment

Once you have configured your environment, you can access the RACE folder and run the code with the following command

python3  src/main.py --config=facmac_smac --env-config=sc2 with env_args.map_name=3s_vs_3z  batch_size_run=1 state_alpha=0.001 frac=0.005  EA_alpha=1.0  Org_alpha=1.0  EA=1  EA_freq=1 SAME=0  use_cuda=False t_max=2005000

state_alpha corresponds to the hyperparameter beta to control VMM, frac corresponds to the hyperparameter alpha to control the mutation, and the other hyperparameters are consistent across tasks. The code is run in serial mode and does not use multi-processing.

The config files src/config/default.yaml contain hyperparameters for the algorithms. These were sometimes changed when running the experiments on different tasks. Please see the Appendix of the paper for the exact hyper-parameters used.

For each environment, you can specify the specific scenario by with env_args.map_name=<map_name> for SMAC.

We found that the PYMARL framework built on SMAC did not work well on MAMUJOCO, so we rebuilt a set of repositories to run MAMUJOCO. The code of MAMUJOCO I will release later, you can also contact me to get this source code in advance

Citing

If you used this code in your research or found it helpful, please consider citing the following paper:

Bibtex:

@InProceedings{pmlr-v202-li23i,
  title = 	 {{RACE}: Improve Multi-Agent Reinforcement Learning with Representation Asymmetry and Collaborative Evolution},
  author =       {Li, Pengyi and Hao, Jianye and Tang, Hongyao and Zheng, Yan and Fu, Xian},
  booktitle = 	 {Proceedings of the 40th International Conference on Machine Learning},
  pages = 	 {19490--19503},
  year = 	 {2023},
  volume = 	 {202},
  publisher =    {PMLR}
}