torin

Results 3 issues of torin

when start a task and load imgs and prelabel files, if the img-label pairs' file number is large, the client would crash. i guess the program preload all annotation at...

### Motivation In downstream tasks, Lora is one of the most common way to finetune llm. The inference speed degrades awfully from [turbomind backend+merge lora ]to [pytorch backend + merge...

awaiting response
Stale

### Motivation vllm has established a V1 version, which supports prefix caching in multi modality llm. As a comparable infer engine,Wish LMDeploy has comparable feature :) ### Related resources _No...

planned feature