LLaMA-Factory
LLaMA-Factory copied to clipboard
FSDP + Qlora Faill
Reminder
- [X] I have read the README and searched the existing issues.
Reproduction
I pulled new code and ran Accelerate +FSDP + Qlora training, but encountered an error: ValueError: Must flatten tensors with uniform dtype but got torch.bfloat16 and torch.float16
However, when I use the old code, it works fine. I suspect there is a bug in the new code.
Expected behavior
No response
System Info
No response
Others
No response