Santhana Mikhail Antony S
Santhana Mikhail Antony S
> see: [#4186 (comment)](https://github.com/vllm-project/vllm/issues/4186#issuecomment-2241307064) But if I merge the LoRA adapter, I cannot use multiple LoRAs, right?
Hey Antony here. I am new to numpy. Will change as many files as possible and raise a PR.
any fix for this issue?
Getting this same error when I try to hit a loaded model from the start with A40 :(
> Ah it seem like this one doesn't have softcapping (#2273). I'd recommend using the latest TGI version. > > Would that work for you? Can you take a look...
> I think this is a `ChatHuggingFace` bug. > > You have to call `bind` on the chat object with the parameters you want changed before running `invoke`. They do...