GPTQ-for-LLaMa
GPTQ-for-LLaMa copied to clipboard
Inference using CPU
Hi, I'm interested to run LLAMA 4bit GPTQ, but I don't have a GPU. Is it possible to run this model on CPU only?