podman-desktop-extension-ai-lab icon indicating copy to clipboard operation
podman-desktop-extension-ai-lab copied to clipboard

ollama serve/proxy support

Open maxandersen opened this issue 1 year ago • 2 comments

Is your feature request related to a problem? Please describe

i see latest nightly has pull and list available like ollama - awesome.

allows me to use ollama list/pull.

any chance to trigger what corresponds to a ollama serve and have /api/generate and/or /api/embeddings work as a proxy so users nor apps don't need to lookup the random generated port number for the running service?

Describe the solution you'd like

have way to run models from api or command line and have a stable host/port to connect with a openai or similar genai serving api relaying to the underlying started container.

Describe alternatives you've considered

No response

Additional context

No response

maxandersen avatar Sep 26 '24 04:09 maxandersen

Are you requesting a new api verb like https://github.com/ollama/ollama/blob/main/docs/api.md#generate-a-chat-completion ?

jeffmaury avatar Sep 26 '24 06:09 jeffmaury

That api is the ollama api - it would be nice-to-have but openai compatible endpoint like https://github.com/ollama/ollama/blob/main/docs/openai.md i.e. v1/chat/completions would be much preferred.

maxandersen avatar Oct 07 '24 12:10 maxandersen