xmc.dspy icon indicating copy to clipboard operation
xmc.dspy copied to clipboard

vLLM Compatibility

Open sidjha1 opened this issue 1 year ago • 1 comments

Hello, I was curious whether it was possible to run models locally via vLLM. The README mentions HF TGI for running local models. Looking through the experimental dspy branch it seems that HF TGI is chosen for the model if an OpenAI model is not provided. Should I modify the experimental branch to add vLLM support or is there another way to run local models on vLLM?

sidjha1 avatar Feb 03 '24 23:02 sidjha1

Ideally DSPy handles all of this, and IReRa just uses whatever LLM you supply. To run with vLLM for now, it is indeed best to change how the models are created in the irera branch on DSPy. I'd need to think of a more scalable way of taking care of model providers long-term.

KarelDO avatar Feb 07 '24 17:02 KarelDO