llama-lora-fine-tuning
llama-lora-fine-tuning copied to clipboard
RuntimeError:CUDA error : out of memory
Two 3060 graphics cards with a total memory of 24GB, why would this error still be reported?