llama.cpp
llama.cpp copied to clipboard
qwen model quantized with AWQ and lora weights
Hello, I'm very new to this repo and just read through the quickstart. I am curious does this repo support qwen model quantized with AWQ which has int4 weights and lora weights trained upon the quantized weight (fp16 I think)?