self-operating-computer icon indicating copy to clipboard operation
self-operating-computer copied to clipboard

Open source large language model support

Open andzejsp opened this issue 1 year ago • 11 comments

Is it possible to run this and point it not at OpenAI but to self hosted large language model to do the thing?

andzejsp avatar Dec 06 '23 15:12 andzejsp

@andzejsp this would be possible. It would likely require some slight changes in prompting and some adjustments in functions to the repo. If someone finds a good provider hosting a model that has a good API and key access then go ahead and add it to the project as a PR

joshbickett avatar Dec 06 '23 16:12 joshbickett

@joshbickett would you consider starting with Ollama?

I created a feature request like this: #35

orkutmuratyilmaz avatar Dec 10 '23 10:12 orkutmuratyilmaz

Yeah, if someone could get a PR of a vision model working locally on the project that'd be great I think

joshbickett avatar Dec 10 '23 17:12 joshbickett

Yeah, if someone could get a PR of a vision model working locally on the project that'd be great I think

Would this work? https://llava-vl.github.io/

https://simonwillison.net/2023/Nov/29/llamafile/

Andy1996247 avatar Dec 10 '23 19:12 Andy1996247

@Andy1996247 it sounds like it may work based on what you mentioned in #101

joshbickett avatar Dec 12 '23 18:12 joshbickett

could this work?

https://github.com/petals-infra/chat.petals.dev#apis

norzog avatar Dec 15 '23 04:12 norzog

Not very familiar with Petals Chat. It may work, but I think Llama.cpp is most promising

joshbickett avatar Dec 15 '23 14:12 joshbickett

@Andy1996247 @orkutmuratyilmaz @norzog wanted to mention that we added support for the Gemini model in case you're interested. Was merged with PR #110

joshbickett avatar Dec 19 '23 18:12 joshbickett

@joshbickett thanks for the update. We're one step closer to open source LLM support🤘🏻

orkutmuratyilmaz avatar Dec 19 '23 19:12 orkutmuratyilmaz

would be cool if ollama was supported https://github.com/jmorganca/ollama Simply pointing to ollama instance, and bobs your uncle. Not sure how it all works, but ollama was no pain to set up and very usable, just one line of command to set it up :).

andzejsp avatar Dec 20 '23 07:12 andzejsp

would be cool if ollama was supported https://github.com/jmorganca/ollama

Simply pointing to ollama instance, and bobs your uncle. Not sure how it all works, but ollama was no pain to set up and very usable, just one line of command to set it up :).

Currently working on LLaVA support through ollama as we speak :)

Obviously accuracy will be low, but I think it'll be great to finally have support for an open sourced model!

michaelhhogue avatar Dec 20 '23 11:12 michaelhhogue

Heads up, I think you should be able to stand up your own OpenAI-compatible API here:

https://llama-cpp-python.readthedocs.io/en/latest/server/#multimodal-models

Then this project can point to your self-hosted API instead of OpenAI.

bpshaver avatar Jan 16 '24 15:01 bpshaver

We now have LLaVa available in the project thanks to a PR from @michaelhhogue!

joshbickett avatar Feb 09 '24 04:02 joshbickett

thanks for the LlaVA support:)

orkutmuratyilmaz avatar Feb 12 '24 10:02 orkutmuratyilmaz