haic0
haic0
HI DevTeam, Could you give me a hand to check this issue, thanks so much! After installed the whl package successfully, follow this guide, cd rtp-llm # For cuda12 environment,...
### Your current environment Machine: NV B200 GPU Docker image: vllm/vllm-openai:v0.9.1 model: deepseek-ai/DeepSeek-R1 CUDA: 12.8 Driver Version: 570.133.20 Command: VLLM_USE_V1=1 vllm serve /models/DeepSeek-R1 --tensor-parallel-size 8 --disable-log-requests --trust-remote-code ### 🐛 Describe...
### System Info Issues: The installation of the latest multi-backend-refactor branch failed in the AMD GPU. While switching to the Rocm-bitsandbytes repo, by using the rocm_enabled_multi_backend branch, the installation was...