ray icon indicating copy to clipboard operation
ray copied to clipboard

multi-agent dreamerv3

Open janetwise opened this issue 9 months ago • 2 comments

Description

Saw the dreamerv3 code in rllib has this error flag on multi-agent not supported currently. Is there any plan to support it in the future? What's the approach to extend to multi-agent dreamerv3?

if self.is_multi_agent():
            raise ValueError("DreamerV3 does NOT support multi-agent setups yet!")

Use case

Mult-agent env and agent learning

janetwise avatar May 13 '24 14:05 janetwise

@janetwise Yes, we want to support all online algorithms in multi-agent mode, so DreamerV3 as well. We are still in the process of moving off-policy algorithms over to the new stack and DreamerV3 comes after these. Until the summit we plan to be ready with this move.

simonsays1980 avatar May 15 '24 11:05 simonsays1980

What’s the technical approach and suggestion if I work on extending dreamer to multiple agents? Will the approach wrapping PPO to multiple agents apply in a similar way?

janetwise avatar May 15 '24 14:05 janetwise

@janetwise This approach (using the MultiEnv) will not work for DreamerV3 unless you are also implementing a multi-agent algorithms.dreamerv3.utils.env_runner.EnvRunner that deals with multi-agent observations and actions similar to our .env.multi_agent_env_runner.MultiAgentEnvRunner.

In the future we want to bring DreamerV3 also onto our default env runners, but this will not happen before the Summit 2024.

simonsays1980 avatar May 31 '24 09:05 simonsays1980