Cao Yu

Results 7 comments of Cao Yu

I've met the same response issue in 7b. But after I changed the `torch` version to 2.1.2 and `transformers` version to 4.40.0. The response shows correctly. ``` Loaded LLaVA model:...

Maybe u could use the same [requirements.txt](https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/requirements.txt) in the main branch.

Just did a quick run, should be ok with si as well. ``` Loaded LLaVA model: lmms-lab/llava-onevision-qwen2-7b-si Loading vision tower: google/siglip-so400m-patch14-384 Model Class: LlavaQwenForCausalLM ["This image is a radar chart...

@zui-jiang I met the same issue, from your logs, it seems before the 300s watchdog timeout, the longest input context would be around 13k. I've increased the watchdog timeout to...

--disable-cuda-graph works, but super slow, like halve the speed, from 40+ tokens/s to 10~20 tokens/s. also if I adjust the --watchdog-timeout 36000, it seems nccl timeout still at 600s.

I'm also using 2x3090. I tried something like this, it will offload some parameters to CPU. > Some parameters are on the meta device because they were offloaded to the...