andreaKim
andreaKim
Hello @loadams Thanks for your replying. ```pip install deepspeed``` works without any errors for me. I am running my script on my master node which has CPU only using slurm...
@loadams There is no reason for doing this. I was just following this [tutorial](https://www.philschmid.de/fine-tune-flan-t5) about finetuning t5 model. I found another way to utilize fused adm just adding ```torch_adam=true``` in...
``` pip uninstall deepspeed DS_ACCELERATOR=cuda pip install deepspeed ds_report ``` produces the result like below ``` [2024-05-24 09:17:18,762] [WARNING] [real_accelerator.py:162:get_accelerator] Setting accelerator to CPU. If you have GPU or other...
I have same issue when I use zero stage 3 with latest deepspeed version. This issue occurs after first evaluation step. How could I fix it?
> 그거 `auto_find_all_linears` 이놈때문에 quantlinear에만 LoRA붙혀서 그래요, 저도 동일한 문제 있는데, 그래서 auto_find_all_linears False주면 또 에러나거든요.... 재귀함수 호출에러인데 이거 호출회수 리미트 풀면 또 이상한에러 납니다.. 어떻게 해야될지 모르겠네요 혹시 성공하셨나요?...
> 뭔가 이해가 잘 안가는데, 참고하신 레퍼런스나 좀 자세히 알려주실 수있으실까요..?ㅠㅠ 로라를 붙힌상태로 gptq 돌리셨단 말씀이신지요?? 이렇게 해도 로라 로드, 언로드 잘 되나요?? 레퍼런스는 따로 없고 저의 목적은 Lora로 LLM을...
@imoneoi Any peft would be proper for OpenChat? I tried fine tuning with many fine-tuned models, however i never saw one successfully fine tuned with lora.
@lewtun Thanks for giving such a good example! I am still curious how to adapt my own dataset that is not on hub. For example, when i try to train...
> DS_ACCELERATOR=cuda ds_report @delock Your recommendation made everything perfect! Thanks for giving nice advice on it! I got same results with you Thanks again :)
@qubvel That's typo in the description! :) that's because I typed filenames by myself (I fixed it in the description)