MARL-code-pytorch
MARL-code-pytorch copied to clipboard
Concise pytorch implements of MARL algorithms, including MAPPO, MADDPG, MATD3, QMIX and VDN.
MARL-code-pytorch
Concise pytorch implements of MARL algorithms, including MAPPO, MADDPG, MATD3, QMIX and VDN.
Requirements
python==3.7.9
numpy==1.19.4
pytorch==1.5.0
tensorboard==0.6.0
gym==0.10.5
Multi-Agent Particle-World Environment(MPE)
SMAC-StarCraft Multi-Agent Challenge
Trainning results
1. MAPPO in MPE (discrete action space)
2. MAPPO in StarCraft II(SMAC)
3. QMIX and VDN in StarCraft II(SMAC)
4. MADDPG and MATD3 in MPE (continuous action space)
Some Details
In order to facilitate switching between discrete action space and continuous action space in MPE environments, we make some small modifications in MPE source code.
1. make_env.py
We add an argument named 'discrete' in 'make_env.py',which is a bool variable.
2. environment.py
We also add an argument named 'discrete' in 'environment.py'.
3. How to create a MPE environment?
If your want to use discrete action space mode, you can use 'env=make_env(scenario_name, discrete=True)'
If your want to use continuous action space mode, you can use 'env=make_env(scenario_name, discrete=False)'