SpQR
SpQR copied to clipboard
Will the function of model saving be realized in the future?
Hi, Thank you for sharing your work. The re-produced perplexity for ptb dataset using your code is not matched with the paper. The reproduced is 27.8, while in the paper...
Anxiously excited about getting access to the inference code!
Hello, I have a question, I currently have a model of the llama series that has been fine-tuned with my own dataset. If I want to SpQR quantize it, do...
failed to find the inference code
Anyway we can test the speedup effect?
Been trying to run quantization for falcon-40b on a box with 8 40Gi A100's but I keep getting CUDA memory errors. The readme states that this should be possible, unless...
I did try running the code from your repository; however, when I attempted to add the --save_safetensors feature, the process was interrupted after performing the evaluation. I didn't encounter any...
Hi @Vahe1994, I have fine-tuned a facebook's nllb model on my custom dataset for language translation. Could you provide a guideline on how to preform SpQR of this fine-tuned model?...
Hi I was wondering if you folks can provide SpQR trained model weights on OpenLLaMA? OpenLLaMA has Apache-2.0 and has reported closer to the original LLaMA’s performance on benchmarks. Thanks....