devika
devika copied to clipboard
Ollama Local model issue after update
after clone yesterday version the local model can't be detected to reply even one or two steps like the previous version 24.04.26 21:38:46: root: ERROR : Inference took too long. Model: OLLAMA, Model ID: llama3 24.04.26 21:38:46: root: INFO : SOCKET inference MESSAGE: {'type': 'error', 'message': 'Inference took too long. Please try again.'} 24.04.26 21:38:46: root: WARNING: Inference failed
Same problem for me! Hopefully this get's fixed soon, I was really excited to try this out.
i am also face the same problem 24.04.27 11:48:22: root: INFO : SOCKET tokens MESSAGE: {'token_usage': 730} Model: mistral, Enum: OLLAMA 24.04.27 11:48:23: root: INFO : SOCKET inference MESSAGE: {'type': 'time', 'elapsed_time': '0.00'} 24.04.27 11:49:24: root: ERROR : Inference took too long. Model: OLLAMA, Model ID: mistral 24.04.27 11:49:24: root: INFO : SOCKET inference MESSAGE: {'type': 'error', 'message': 'Inference took too long. Please try again.'} 24.04.27 11:49:24: root: WARNING: Inference failed
Also for me with a 3090 Inference took too long I just use a 20GB command-r verison, normal ollama chat works superfast
Same here.
now you can update the inference timeout via the settings page. fetch latest changes.