continue
continue copied to clipboard
Preload model for the ollama provider
Validations
- [ ] I believe this is a way to improve. I'll try to join the Continue Discord for questions
- [X] I'm not able to find an open issue that requests the same enhancement
Problem
I've noticed that when I use ollama to chat, the model loading always occurs during the first round of conversation, which makes the first round much slower than subsequent ones.
I'm exploring ways to help ollama preload the model. Even though I tried ollama run llama2:latest before conversation, the model still loads at the start of the first conversation.
Solution
No response
@sgwhat https://github.com/ollama/ollama/blob/main/docs/api.md#load-a-model
This might be a good solution
@sgwhat https://github.com/ollama/ollama/blob/main/docs/api.md#load-a-model
This might be a good solution
I have tired, but still useless.