OpenLLM icon indicating copy to clipboard operation
OpenLLM copied to clipboard

Deprecation Warning for PyTorch Backend

Open byteshiva opened this issue 11 months ago • 2 comments

Description

Running the command docker run --rm -it -p 3000:3000 ghcr.io/bentoml/openllm start facebook/opt-1.3b --backend pt results in a deprecation warning for the PyTorch backend.

Error

Error: PyTorch backend is deprecated and will be removed. Switch to vLLM.

Can you let me know how to fix it.

Resolution

  • Update documentation to reflect deprecation.
### Tasks
- [x] Update documentation to reflect the deprecation of the PyTorch backend and provide guidance on transitioning to vLLM or another supported backend
- [x] Modify the Docker image to remove dependencies on the deprecated PyTorch backend
- [x] Test the modified Docker image to ensure it runs without any deprecation warnings.

byteshiva avatar Mar 19 '24 15:03 byteshiva

found the fix

docker run --rm -it -p 3000:3000 ghcr.io/bentoml/openllm start facebook/opt-1.3b --backend vllm

Fix Available at:

  • [x] https://github.com/bentoml/OpenLLM?tab=readme-ov-file#-runtime-implementations
  • [x] ex. openllm start meta-llama/Llama-2-7b-chat-hf --backend vllm

byteshiva avatar Mar 19 '24 15:03 byteshiva

just a headsup for 0.5 I'm removing pytorch backend, will work on CPU specific branch for better optimization.

aarnphm avatar Mar 21 '24 06:03 aarnphm