alpaca-lora icon indicating copy to clipboard operation
alpaca-lora copied to clipboard

Load_in_8bit causing issues: Out of memory error with 44Gb VRAM in my GPU or device_map error

Open Nimisha-Pabbichetty opened this issue 8 months ago • 1 comments

I'm able to get the generate.py script working. However, with the finetune.py script I'm facing the following error: image

It seems to be because the load_in_8bit parameter is set to True and it's looking for a quantisation_config,json but if I set it to False then even a GPU with vRAM of 44Gb is not enough to train the model. How do I create the quantisation_config,json? I'm using huggyllama/llama-7b as the base model since the given link for the base model is down. I face the same error when I use baffo32/decapoda-research-llama-7B-hf as the base model.

Any help would be appreciated, thank you!

Nimisha-Pabbichetty avatar Nov 11 '23 19:11 Nimisha-Pabbichetty