memU icon indicating copy to clipboard operation
memU copied to clipboard

Support for any OpenAI compatible endpoint

Open Awwtifishal opened this issue 4 months ago • 3 comments

All supported providers ultimately use the same API (OpenAI compatible), so why not just support one single API with a choice for the provider endpoint (base URL)?

Also, I would like to choose the endpoint for LLM and for embeddings separately, to be able to use llama.cpp for example (which can only load one model).

Awwtifishal avatar Aug 13 '25 17:08 Awwtifishal

Good point! We will put it into our roadmap and publish memU implemented llm endpoint soon! Will notify when we finished. Thank you for your feedback

sairin1202 avatar Aug 14 '25 05:08 sairin1202

Looking at this commit I'm not sure if you understood my message. All we need is just that, but for embedding instead of the LLM. Just an env variable for the endpoint (base URL) for the OpenAI embedding API which is different than OPENAI_BASE_URL. For example something like OPENAI_EMBEDDING_BASE_URL, which falls back to OPENAI_BASE_URL for the people that don't need a different endpoint.

Awwtifishal avatar Aug 18 '25 21:08 Awwtifishal

I managed to make a proxy that can serve both the LLM and embeddings from the same endpoint (openai compatible), and it seems that the server tests work.

I'm not closing the issue because of other folks that want to try a local deployment.

Awwtifishal avatar Aug 22 '25 20:08 Awwtifishal