twinny
twinny copied to clipboard
Improvement limit prompt length to LLM context window
- Adhere to LLM context window size when conversations or fim context overflow the limit.
I glanced at the Ollama API and wondering, is there a way from various backends to get that information from them?
https://github.com/ollama/ollama/blob/main/docs/api.md