Andrew Phillips
Andrew Phillips
@manojm-dev Sorry, I hadn't noticed this PR. I would prefer if this didn't change the default port, could you make it use a `PORT` env variable instead?
I guess this is the same problem as in #5022.
@sammcj, I don't think your patch fixes this, but now I have a larger context anyway. This PR seems to indicate there is some double counting going on in the...
@rick-github Good point about `num_gpu` (note that `ollama ps` will still say it's splitting it even if it doesn't), however I think the estimate is also used when deciding to...
Are we sure the ollama ps output is in base 10? From my earlier comment: ``` NAME ID SIZE PROCESSOR UNTIL DEFAULT/mistral-small-2409-22b:latest d9db479f49e8 24 GB 100% GPU Forever ``` I...
Sorry, yeah that's painfully obvious now that I'm re-reading it later. You are correct.
FYI, I made a PR to add `ollama ps --base2`: https://github.com/ollama/ollama/pull/8034 ``` industrial:~/projects/ollama-src$ ./ollama ps --base2 NAME ID SIZE PROCESSOR UNTIL DEFAULT/mistral-small-2409-22b:latest 671ad04c21ce 24.4 GiB 7%/93% CPU/GPU Forever industrial:~/projects/ollama-src$ nvidia-smi...
> @theasp pr #8034 doesn't fix how Ollama overestimates memory usage and offloads incorrect number of layers, right? > Running `ollama ps --base2` just offsets and shows numbers close to...
Hi Peter, No urgency on my part, take your time. I didn't even notice for over a year!
In my case, I can't run `opencode debug config` without unsetting `http_proxy` and `https_proxy`. If I do unset them it prints out a valid config.