OpenLLM icon indicating copy to clipboard operation
OpenLLM copied to clipboard

feat: c4ai-command-r-v01 support

Open 0x77dev opened this issue 1 year ago • 2 comments

Feature request

Would be nice to have ability to run Command-R (CohereForAI/c4ai-command-r-v01) using OpenLLM

Motivation

No response

Other

vLLM backend already supports Command-R in v0.4.0: https://github.com/vllm-project/vllm/issues/3330#issuecomment-2041225404

0x77dev avatar Apr 06 '24 23:04 0x77dev

The current ghcr.io/bentoml/openllm:latest image (sha256:1860863091163a8e8cb1225c99d6e1b0735c11871e14e8d8424a22a5ad6742fa) shows an error:

ValueError: The checkpoint you are trying to load has a model type of `cohere`, which Transformers does not recognize. This may be due to a problem with the checkpoint or an outdated version of Transformers.

when doing this:

docker run --rm --gpus all -p 3000:3000 -it ghcr.io/bentoml/openllm start CohereForAI/c4ai-command-r-v01 --backend vllm

also when installing openllm[vllm] it brings 0.2.7 version of vLLM

Though vLLM version in main branch is 0.4.0: https://github.com/bentoml/OpenLLM/blob/main/openllm-core/pyproject.toml#L83 and https://github.com/bentoml/OpenLLM/blob/main/tools/dependencies.py#L157

0x77dev avatar Apr 06 '24 23:04 0x77dev

I think this should be the same prompting system, there is also CohereForAI/c4ai-command-r-plus available and it would be nice to be able to run it too.

GaetanBaert avatar Apr 10 '24 07:04 GaetanBaert

should be supported on main now. Will release a new version soon.

aarnphm avatar Jun 03 '24 17:06 aarnphm