OpenLLM icon indicating copy to clipboard operation
OpenLLM copied to clipboard

PEFT LORA / QLORA

Open louisoutin opened this issue 2 years ago • 1 comments

Hello, Are you planning to add support for parameter efficient finetuning methods? Also does it support doing inference using those adapters models to optimize VRAM? Thanks

louisoutin avatar Jun 20 '23 16:06 louisoutin

Yes, I'm currently working on fine tuning API atm

aarnphm avatar Jun 20 '23 16:06 aarnphm

Peft and LoRA is now supported during serving and building time via --adapter-id

I will continue to gradually add support for all default fine-tuning layers for each of the models.

Will create a new tracking issue to specific set method support for each of these models

aarnphm avatar Jun 24 '23 08:06 aarnphm