Extremys
Extremys
Hello, Any update on this feature?
Do you know if it has been fixed since so far? with v1.40.3 I still have the issue FYI
Great! For my part, to be exact, I'm trying to use a [TEI](https://github.com/huggingface/text-embeddings-inference) instance but it should be pretty similar. We have the swagger there https://huggingface.github.io/text-embeddings-inference/ There is also /v1/embeddings...
> what's the error you're seeing? @Extremys > > wondering why `openai/` doesn't just work https://docs.litellm.ai/docs/providers/openai_compatible#usage---embedding I'm getting this: ```log 15:30:26 - LiteLLM:INFO: utils.py:1310 - POST Request Sent from LiteLLM:...
> @Extremys can you bump and re-share. We now show the received error from provider in the 'no deployments available' message It should be better now :)
> > RAW RESPONSE: > > AsyncEmbeddings.create() got an unexpected keyword argument 'no_proxy' > > @Extremys it looks like the no_proxy keyword is causing errors. Is that intended? Not really,...
Yes amazing! It would be so great to have also OpenAI-like API compatibility, so many Open sources multimodal models are available like Idefics2, Llava, llama-3-vision, ... :)
> I used chat-ui-db (https://github.com/huggingface/chat-ui/pkgs/container/chat-ui-db) as the front-end and vllm (https://github.com/vllm-project/vllm) as the back-end for large language model inference. Llama-2-70b-chat-hf model weights used. The content of the .env.local file is...