litgpt
litgpt copied to clipboard
Phi-3 Full finetuning uses more memory as LoRA finetuning
I observed that Phi-3 full finetuning uses less memory than LoRA (see #1553); as discussed something to look into @Andrei-Aksionov