frob
frob
You will need to provide more information to debug this: the system prompt you are using, a sample query that doesn't return the correct results, and ideally a capture of...
https://github.com/ollama/ollama/issues/5965#issuecomment-2252354726
[Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) will help in debugging.
Yes, but there are no relevant server logs. Posting one `Debug line of interest:` with no context is insufficient. If there's a bug somewhere, having information on how ollama and...
Thanks, let me dig in to this.
OK, I think there is a misunderstanding of how ollama manages the context size. The value of `llama.context_length` in the model parameter is the maximum context window that the model...
Once again, server logs of the failure would be illuminating.
OK, I see from `POST "/v1/chat/completions"` that you are using the OpenAI API comparability endpoints. The OpenAI API standard doesn't support setting the size of the context window. The only...
`num_ctx` in OpenAI endpoints (`localhost:11434/v1`) are ignored. If you send a request with `num_ctx` to the ollama endpoints (`localhost:11434/api`) and it's different to the the value of `num_ctx` in the...
You can file a ticket to have the Modelfile updated, but I'm not sure if it's a good idea. Memory usage scales by the size of the context window, llama3.1...