frob

Results 772 comments of frob

You will need to provide more information to debug this: the system prompt you are using, a sample query that doesn't return the correct results, and ideally a capture of...

https://github.com/ollama/ollama/issues/5965#issuecomment-2252354726

[Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) will help in debugging.

Yes, but there are no relevant server logs. Posting one `Debug line of interest:` with no context is insufficient. If there's a bug somewhere, having information on how ollama and...

OK, I think there is a misunderstanding of how ollama manages the context size. The value of `llama.context_length` in the model parameter is the maximum context window that the model...

Once again, server logs of the failure would be illuminating.

OK, I see from `POST "/v1/chat/completions"` that you are using the OpenAI API comparability endpoints. The OpenAI API standard doesn't support setting the size of the context window. The only...

`num_ctx` in OpenAI endpoints (`localhost:11434/v1`) are ignored. If you send a request with `num_ctx` to the ollama endpoints (`localhost:11434/api`) and it's different to the the value of `num_ctx` in the...

You can file a ticket to have the Modelfile updated, but I'm not sure if it's a good idea. Memory usage scales by the size of the context window, llama3.1...