ollama
ollama copied to clipboard
Enable Prompt Caching by Default
I had to disable prompt caching due to requests getting stuck: #1994
We should bring this back when we have a mitigation for the inference issue: https://github.com/ggerganov/llama.cpp/issues/4989