llama.cpp
llama.cpp copied to clipboard
Qwen2.5-vl support and conversion?
Hi,big guys! I would like to ask if llama.cpp is able to convert a multimodal model (e.g. Qwen2.5-vl-3B) to gguf format and quantize it to Q4_0? And is there a corresponding tool to run it through llama.cpp after compilation? If not, are there any subsequent plans for this part of the development?
See: https://github.com/ggml-org/llama.cpp/issues/11483#issuecomment-2727577078
This issue was closed because it has been inactive for 14 days since being marked as stale.