frob

Results 840 comments of frob

Server logs would be helpful to diagnose the issue.

Not all layers are being loaded onto the GPU in 0.2.7: ``` llm_load_tensors: offloading 41 repeating layers to GPU llm_load_tensors: offloaded 41/43 layers to GPU ``` Do you have corresponding...

Not all 8G on your card is available for ollama: `memory.available="[6.8 GiB]"`. What's the output of `nvidia-smi`?

Does it change if you turn flash attention off? My understanding is that ollama does the calculations for how many layers to offload based on the size of the model...

You can try, but they will also probably ask you for the output of `nvidia-smi`.

It's unfortunate that `nvidia-smi` on windows doesn't report GPU Memory Usage. But we see here that your Brave browser is taking up GPU memory. If Brave works like other browsers,...

[Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) may aid in debugging.

If you could add the full log that would help, there are lines earlier with information about device detection, environment, etc. It may also help if you add `OLLAMA_DEBUG=1` to...

[Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) may aid in debugging.