zzb213213
zzb213213
同问,相同的报错,InternVL3-8B就可以直接量化
2025-11-15 11:59:12,743 - lmdeploy - WARNING - archs.py:45 - Fallback to pytorch engine because `/model/darnellzhu/Qwen-25-VL/Qwen3-VL-8B-sft` not supported by turbomind engine. huggingface/tokenizers: The current process just got forked, after parallelism has...
> 请使用此提交:[02cd79b](https://github.com/InternLM/lmdeploy/commit/02cd79b6a3f9105d3a87fabb26b49b40bc084975) 采用llama_factory微调qwen3-vl,llama_factory推理结果没问题,lmdeploy推理结果完全不正确,输出格式正确