fsdp_qlora
fsdp_qlora copied to clipboard
Request for Scripts to Merge QDoRA Adapters with Base Model for vLLM Inference
Hello,
I've successfully finetuned Llama-3 8B with QDoRA and am now looking to perform inference using vLLM. Could you provide guidance or scripts on how to merge the QDoRA adapters with the original base model? Additionally, does this process involve quantization and dequantization of the base model?
Thank you!