Ettore Di Giacinto
Ettore Di Giacinto
@malanthrax could you share your full logs with the `--debug` flag?

I have a bit of mixed feeling in this, my thoughts inline: > Part of the CI Speedup Project https://github.com/users/dave-gray101/projects/2/views/1 > > In order to _really_ accelerate CI, I want...
cc @lu-zero
is that a gguf model? which backend is being used? the logs are not enough here to debug properly. Can you give the full snippet?
@billy-sung here gpt-oss seems to work fine. What version of LocalAI are you running? it's visible in the footer of the webUI.
@billy-sung is that a new install? Did you tried to pull again the image? you need the latest LocalAI due to the new template format (we are at `3.4.0`)
Currently there are two ways to make LocalAI unload memory during runtime: - By setting `LOCALAI_SINGLE_ACTIVE_BACKEND`, which will make sure there is only one model loaded - By setting `LOCALAI_WATCHDOG_IDLE=true`...
> merged and squashed this again. > > @mudler - Can you decide if this PR is worth maintaining? It's been stable through my testing for a while now. I'd...
> @jimmykarily When are we expecting the code, can I take a week's time? Sure thing @HariniKrishnan ! Thank you :100: