jaszhu13

Results 1 issues of jaszhu13

In qlora.py line https://github.com/artidoro/qlora/blob/main/qlora.py#L279 , if fp16 is specified we assign torch_dtype to torch.float32? Shall we do torch.float16 instead, or this is intentional, if so what's the reason and why...