frob

Results 849 comments of frob

From the logs, Flash Attention is enabled in 0.12.10 but not 0.12.5. Try setting `OLLAMA_FLASH_ATTENTION=0` in the server environment.

https://github.com/ggerganov/llama.cpp/pull/8967

[Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) will aid in debugging.

``` ollama pull hf.co/Qwen/Qwen3-Embedding-0.6B-GGUF:Q8_0 ``` ollama doesn't currently support ranking models, #3368.

Just import it as show in https://github.com/ollama/ollama/issues/10989#issuecomment-2946454983.

No. https://arxiv.org/abs/1708.03629

A model needs to have a `pooling_type` field in the KV metadata in order to have `embedding` as a listed capability. That's a part of the GGUF file, there's no...

qwen3 embedding is already supported.