Multi-Agent-Deep-Deterministic-Policy-Gradients icon indicating copy to clipboard operation
Multi-Agent-Deep-Deterministic-Policy-Gradients copied to clipboard

A Pytorch implementation of the multi agent deep deterministic policy gradients (MADDPG) algorithm

Results 9 Multi-Agent-Deep-Deterministic-Policy-Gradients issues
Sort by recently updated
recently updated
newest added

I tried to run simply_reference scenario but this error appeared. AttributeError: 'MultiDiscrete' object has no attribute 'n' How to fix it please

https://github.com/philtabor/Multi-Agent-Deep-Deterministic-Policy-Gradients/blob/a3c294aa6834f348a7401306dff3e67919c861f5/maddpg.py#L74 Hi Phill, Could you please help me to understand what's this line is for? critic_value_new[dones[:, 0.0]] = 0.0 Since critic_value_new float variable it cannot be used as array. Should...

https://github.com/philtabor/Multi-Agent-Deep-Deterministic-Policy-Gradients/blob/a3c294aa6834f348a7401306dff3e67919c861f5/maddpg.py#L83 Dear Phill, First of all plenty of thanks and gratitude for your lessens, I've learned a lot from your lectures. I've noticed a difference in the code at line...

**After running, I get the following error. Where does the problem come from?** No such file or directory: 'tmp/maddpg/simple_adversary/agent_0_actor'

here is the solution https://github.com/philtabor/Multi-Agent-Deep-Deterministic-Policy-Gradients/issues/2#issuecomment-912548033

`self.target_critic.load_state_dict(critic_state_dict)` above code seems make target critic network's parameter always be same as the critic network's. So what is the purpose? making the network learn more slowly? Hope somebody help...

When I run the code, the following error appears, showing that the make_env module is missing,how to solve this problem? Traceback (most recent call last): File "F:\project\pytorch\edge computing\Multi-Agent-Deep-Deterministic-Policy-Gradients-master\main.py", line 4,...

File "maddpg_torch.py", line 345, in memory.store_transition(obs, state, actions, reward, obs_, state_, done) File "maddpg_torch.py", line 51, in store_transition self.state_memory[index] = state ValueError: could not broadcast input array from shape (8)...

重构基于pettingzoo MPE环境的MADDPG及MATD3算法,且重载mpe环境,支持自定义环境! 解决若干问题,如渲染问题等。 请查看我的仓库:https://github.com/Ronchy2000/Multi-agent-RL/tree/master/MADDPG_Continous 任何问题欢迎提PR和issue! --- Refactor the MADDPG and MATD3 algorithms based on the pettingzoo MPE environment, and overload the mpe environment to support custom environments! Solve some problems,...