UMOE-Scaling-Unified-Multimodal-LLMs
UMOE-Scaling-Unified-Multimodal-LLMs copied to clipboard
Clarification on 3-Step Training Approach and Commands for Uni-MoE v2
I like the three step innovative training approach to train the MLLMs. This intrigued me more and I was going through the scripts trying to replicate 3 step training technique to train my own model. However, I have few queries.
- is it possible to replicate all three training steps with the scripts in uni-moe-v2 folder?
- Could you share the command to train uni-moe-v2-speech as there are only inference and eval scripts?
- relating to the 3 step training approach and the given model checkpoints, Uni-MoE 8-expert base is the result of step1, Uni_MoE 8-expert experts model after step 2 and Uni_MoE 8-expert finetune model is the model after step 3. Is my understanding correct?