How to train a multi-agent video understanding task
I have read your paper, which have great help for my work. In your paper, you said that agent-lightning supports multi-agent reinforcement learning with any agent and almost no code change. However, I can not read how to train a multi-agent (eg. 2 Qwen2.5-VL-7B training at the same time)system in the readme and guidance. Can you add the content about how to train a multi-agent framework in video-understanding?
We can't train two models at the same time due to a verl limitation.
We can however train two agents alternatively by specifying the trained_agents parameter in LitAgent.
Do you have any examples or guidance about training 2 agents alternatively?