Anton

Results 52 comments of Anton

Tried some 14b models for code on my homelab server, but saw a comment they are too small. I rented a server on vast ai, and neither `devstral:24b` nor `llama3.1:70b`...

I think, the reason, is the client sends num_ctx in a wrong manner, therefore ollama defaults to 4k. I set default context in `OLLAMA_CONTEXT_LENGTH` env var of the ollama systemd...