maddpg icon indicating copy to clipboard operation
maddpg copied to clipboard

Code for the MADDPG algorithm from the paper "Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments"

Results 48 maddpg issues
Sort by recently updated
recently updated
newest added

Running the train.py using the simple crypto scenario gives me this error, I have tried printing out the value of the alpha but all I got was only one value...

Hello,This happens in experiments: Using good policy maddpg and adv policy maddpg Loading previous state... What is the meaning?

Hi, I dont understand 'u_action_space = spaces.Discrete(world.dim_p * 2 + 1)' I know that action[0] is the communication, but why dim_p needs to mutiply 2

I'm trying to set the random seeds so I can get a unique result several times. tried setting np seed in several scripts but none of them give the what...

I have two problem about update function after read your code. And could anyone explanation it for me? I am very appreciated. Firstly, I can't understand what role does variable...

Hello, I am very interested in this project. However, although the work is almost perfect, I think there still be some amendment to this. Please refer to the line in...

I don't really follow the way how you update the actor. From my understanding, the chain rule is required for the gradient of the parameters in the actor, right? But,...

Hello, I wanted to verify something I found in your code. In `maddpg.py` , there seems to be some problem with `obs[None]` in `self.action`. The error reported by the ide...

I typed the policy as 'lol' and it still gives the same result. The readme says that we can run the experiments for ddpg as well as maddpg which is...

Hello, I wanted to verify something I found in your code. In the method `MADDPGAgentTrainer.update()` there is a comment next to the following line stating that an update is only...