Ettore Di Giacinto
Ettore Di Giacinto
Is this still relevant?
> **Description** > > This PR fixes #7189 > > **Notes for Reviewers** > > After applying the fix I see the right reranker output: > > ``` > curl...
Thanks for filing the issue, this is something indeed we should support, either ootb or explicitly opting-out from the cloud config.
did you check the binary releases? https://github.com/mudler/LocalAI/releases/tag/v2.18.1
Can confirm this, fix is on its way in #7235
> #7025 introduced handling result cropping by `top_n`. However, I believe the most users just omit `top_n`, just because don't bother to count `len(documents)`. So, I'm afraid releasing #7025 causes...
Hey @titogrima - LocalAI doesn't set any thread when running in `p2p` mode. This sounds more like a bug in `llama.cpp` as we just run the vanilla rpc service from...
Also might be worth noting that you can pass any command options of llama.cpp from LocalAI with `--llama-cpp-args` or `LOCALAI_EXTRA_LLAMA_CPP_ARGS`, from the `--help` output: ``` ./local-ai worker p2p-llama-cpp-rpc --help Usage:...
mmh seems libwhisper.a was removed from Make targets, but bindings still refer to it :https://github.com/ggerganov/whisper.cpp/blob/37c88027e1cb7c4d8f9c7bc874f7da6d3e179984/bindings/go/Makefile#L33
Hey, thanks for looking at this! > presumably it might be fixed for llama_cpp rerank as well, if @mudler consider it necessary. I think for llama.cpp we should be covered...