Wenxuan Tan

Results 46 comments of Wenxuan Tan

Hi, It's on our schedule, maybe in two weeks or so

Thanks for the issue. Could you try running that again? I pushed a fix a few weeks back

Both @ver217 and I have seen this bug, which appears when pp is off. Will dig more into it

Hi, Could you try pulling the latest main branch? I don't have trouble running pp_size = 2.

I think the booster should support any dataset. Have you tried replacing the random dataset with this? https://github.com/hpcaitech/ColossalAI/blob/8020f4263095373e4c7ad1b15e54b966a8ccb683/examples/language/llama2/finetune.py#L209

Actually, in pp only the last stage computes loss, so this is not a bug. You'll need to do this to see the actual loss. Also, there's a llama fine-tuning...