Tung Son Tran
Tung Son Tran
Have anyone succeeded with porting this to GPT-Neo?
I'm also having the same issue, not just task 2 but any sub-task of task 1 and even the pretrain 2017 task. Can anyone please help on this?
I also am having this problem, but it was caused by trainer.train()
This problem still persists even after updating to the latest version > Hi everyone, not sure if this is related but this might be fixed on peft main: [huggingface/transformers#28770 (comment)](https://github.com/huggingface/transformers/issues/28770#issuecomment-1935819776)...
The problem seems to arise when using Peft+Deepspeed, even when training on only 1 GPU. My code would have no problem when running without distributed training. ``` home/tung/anaconda3/envs/gpt/lib/python3.9/site-packages/torch/nn/modules/module.py:1879: UserWarning: Positional...
> Was anyone able to find a solution to this problem? I'm also not able to resume from checkpoint, using deepspeed zero 3 Do you use DS zero 3 with...