auto-code-rover icon indicating copy to clipboard operation
auto-code-rover copied to clipboard

Adding a support for Cohere Command-R, Anthropic Claude, and Gemini APIs

Open adam2am opened this issue 10 months ago • 7 comments

Hey, I would like to suggest support for integrating additional language model APIs beyond just OpenAI. Specifically, it would be very helpful to have the ability to use:

  • Cohere API (including recent Command-R model with amazing retrieval-augmented generation)
  • Anthropic API (Claude model)
  • Google Gemini API
  • Ollama Local LLMs (for the sake of those, who can't share the code cuz of the NDA)

These models rank among the top 10 AI language models according to benchmarks like https://chat.lmsys.org/ and provide capabilities complementary to OpenAI's models.

The recent Command-R model from Cohere is particularly compelling for its strong retrieval-augmented capabilities using its embeddings. And the Claude model from Anthropic has received acclaim for its coherence and abilities to code.

Having this flexibility would be incredibly valuable. Would be amazing if you consider adding it!

adam2am avatar Apr 12 '24 15:04 adam2am

It would be good to add AWS Bedrock API to that list. It supports Claude, LLama, etc...

IvoP1 avatar Apr 13 '24 19:04 IvoP1

Just add litellm support. Everything already included there. No need to add support for each and every LLM and provider

andr0s avatar Apr 21 '24 03:04 andr0s

We are currently working on the LiteLLM support. It should be ready sometime next week.

yuntongzhang avatar Apr 21 '24 03:04 yuntongzhang

Just add litellm support. Everything already included there. No need to add support for each and every LLM and provider

LiteLLM is a great project, but some models have no native support for function calling , etc. Amazon Bedrock Claude3

stevensu1977 avatar Apr 28 '24 03:04 stevensu1977

@stevensu1977 well, the whole point is to have the ability to run open source and/or free things. Or experiment with different LLMS. For example, I'm really curios about Llama 3 70b and that big llm from databrix - in my tests on other projects, they showed pretty decent results, despite being dirt cheap or even free (depending on where you run them). I think others also willing to experiment or reduce costs, otherwise people would just be happy with ChatGPT and Claude 3, and this ticket wouldn't even exist. Another problem with close source models that I noticed - sometimes they change things and it just breaks everything. I'm using Cursor IDE with ChatGPT and the same model suddenly started to perform much worse, basically breaking the whole code, removing tons of my code and adding stubs like "... unchanged code... " which was extremely annoying. I think it wouldn't happen if the model and the backend layer above it would be exactly the same, which is guaranteed only when using open source solutions

andr0s avatar Apr 28 '24 04:04 andr0s

ACR now supports Claude models through LiteLLM and Llama3 through Ollama (added in https://github.com/nus-apr/auto-code-rover/pull/32). Feel free to try it out :)

yuntongzhang avatar Apr 30 '24 12:04 yuntongzhang

can you please add support for all open LLM through liteLLM or ollama? currently, for code generation, codeQwen 1.5 7b is ranked as high as GPT-4 and it's only a 7b model so it can easily run locally pretty quickly.

rezzie-rich avatar May 03 '24 03:05 rezzie-rich

Hi @rezzie-rich, justa added support for generic litellm in commit bed6e775bdc701ca8c0242e072c5c2ac25098477 Ollama's been supported for a while now, will close the issue

Marti2203 avatar Jul 18 '24 16:07 Marti2203