podman-desktop-extension-ai-lab
podman-desktop-extension-ai-lab copied to clipboard
ollama serve/proxy support
Is your feature request related to a problem? Please describe
i see latest nightly has pull and list available like ollama - awesome.
allows me to use ollama list/pull.
any chance to trigger what corresponds to a ollama serve and have /api/generate and/or /api/embeddings work as a proxy so users nor apps don't need to lookup the random generated port number for the running service?
Describe the solution you'd like
have way to run models from api or command line and have a stable host/port to connect with a openai or similar genai serving api relaying to the underlying started container.
Describe alternatives you've considered
No response
Additional context
No response
Are you requesting a new api verb like https://github.com/ollama/ollama/blob/main/docs/api.md#generate-a-chat-completion ?
That api is the ollama api - it would be nice-to-have but openai compatible endpoint like https://github.com/ollama/ollama/blob/main/docs/openai.md i.e. v1/chat/completions would be much preferred.