Possible unsupported configuration: using Ollama for embedding model
Description
I'm using Open WebUI with the confluence search tool and and am trying to use an embedding model using Ollama. However, when I run the tool, it attempts to download the model from Hugging Face instead.
I'm not sure if Ollama is a supported source for embedding models, or if I'm misconfiguring something. I'd appreciate clarification on whether this use case is supported.
Steps to Reproduce
Ollama and Open WebUI are running and work properly.
I set the valve options to snowflake-arctic-embed2:latest and tried multiple combinations to specify the path to the Ollama model path.
I also set the RAG_EMBEDDING_MODEL variable correspondingly.
Expected Behavior
I expected the tool to detect or use the embedding model provided via Ollama, avoiding the need to download anything from Hugging Face.
Actual Behavior
The tool initiates a model download from Hugging Face which fails every time, since the server is offline and is maintained locally.
Environment
OS: 6.14.2-arch1-1
Python version: 3.13.3
Open WebUI version: 0.6.0
Ollama version: 0.6.6
Additional Context
I searched the documentation and issues but couldn't find a clear statement on Ollama compatibility for embedding models.
If this isn't supported yet, I'd be curious to know if it's planned or how best to contribute toward adding it.
Hey, thanks for pointing out that Ollama embedding models aren't supported yet. It's definitely doable, and I think it would be a good addition.
If you're interested in contributing to this, help is welcome. We won't be supporting it officially just yet, but we can discuss the implementation details and make sure it integrates well with the tool.