Local LLMs Support
Hi, thanks for building and opening Savvy!
Is there any way I can configure it to use a locally-running LLM? With OpenAI-compatible API or otherwise.
Thanks!
Curious about your use case for local LLMs vs OpenAI ?
Usage in no-network conditions, data protection, ability to choose specific models for specific kind of workload (for example fine-tuned)
Nothing unique, just a "local LLM" use-case
Thanks for sharing!
Local Redaction and support for local LLMs is planned and I'm tracking it on our public feedback board here
Hi @av,
quick update:
I've moved away from using OpenAI for generating runbooks and we now use Llama3.1 hosted on Groq.
Savvy ask/explain still uses GPT4-o for now.
I'm interested in this too. Would be great if it respected something like OPENAI_API_BASE for the backend to query. Just poked through the code (for about 10 seconds, so take it with a grain of salt) and it looks like all the prompting etc is done server-side, so this would require replicating that logic here.
@sjuxax and @av getting started on this - should have something for y'all to try v soon.
@sjuxax and @av you can now BYO LLM with Savvy's CLI. See the docs here: https://docs.getsavvy.so/guides/byo_llm/
Implemented in #154