worker-vllm icon indicating copy to clipboard operation
worker-vllm copied to clipboard

Are visual LLMs supported?

Open iamkhalidbashir opened this issue 1 year ago • 6 comments

I am trying to run https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct on serverless but it throws error

2024-10-09 18:06:06.525
[my1jfmdyltrax5]
[error]
worker exited with exit code 1
2024-10-09 18:05:50.528
[my1jfmdyltrax5]
[error]
worker exited with exit code 1
2024-10-09 18:05:34.515
[my1jfmdyltrax5]
[error]
worker exited with exit code 1
2024-10-09 18:05:19.851
[my1jfmdyltrax5]
[error]
worker exited with exit code 1
2024-10-09 18:04:18.168
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:04:01.663
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:03:45.158
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:03:28.734
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:03:12.261
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:02:55.743
[eau4o046d9uc4d]
[error]
worker exited with exit code 1
2024-10-09 18:02:39.224
[eau4o046d9uc4d]
[error]
worker exited with exit code 1

iamkhalidbashir avatar Oct 09 '24 13:10 iamkhalidbashir

multi-modal models like Qwen2-VL is supported in pods, but it seems not in serverless endpoints. see my comment here: https://github.com/runpod-workers/worker-vllm/issues/114#issuecomment-2412967748

cris-almodovar avatar Oct 15 '24 06:10 cris-almodovar

yes, this is fixed in the dev branch of vllm which when released in the next version should be fixed for endpoints as well.

iamkhalidbashir avatar Oct 15 '24 06:10 iamkhalidbashir

Hi there!, At the moment we do not support visual-LLMs, but we are working on support for these in the future

therealadityashankar avatar Oct 15 '24 09:10 therealadityashankar

Hey! Any update about visual-llms support?

edtjulien avatar Dec 03 '24 19:12 edtjulien

It is supported now!

nerdylive123 avatar May 31 '25 01:05 nerdylive123

It is supported now!

what model did you manage to run? I've tried multiple ones but none seems to work with the current vllm-worker

faleandroid avatar Aug 05 '25 10:08 faleandroid