Wenxuan Tan
Wenxuan Tan
killing remote host sometimes works
Hi, It's on our schedule, maybe in two weeks or so
I will take multiple looks
I plan to release it next week
Thanks for the issue. Could you try running that again? I pushed a fix a few weeks back
@ver217 any insights?
Both @ver217 and I have seen this bug, which appears when pp is off. Will dig more into it
Hi, Could you try pulling the latest main branch? I don't have trouble running pp_size = 2.
I think the booster should support any dataset. Have you tried replacing the random dataset with this? https://github.com/hpcaitech/ColossalAI/blob/8020f4263095373e4c7ad1b15e54b966a8ccb683/examples/language/llama2/finetune.py#L209
Actually, in pp only the last stage computes loss, so this is not a bug. You'll need to do this to see the actual loss. Also, there's a llama fine-tuning...