InternVL
InternVL copied to clipboard
Convert to Gguf format to work with Llama.cpp?
Llava has various quantized models in gguf format, so it can be used with Llama.cpp. https://github.com/ggerganov/llama.cpp/pull/3436 Is this possible?
Hi, thank you for your suggestion. I will add compatibility with community tools to my to-do list.
gguf format is good for ollama users, Any Update?
It will be nice to have this model in gguf format in ollama.
any updates on this? the 4b intern model is killer for its size! would love to see it supported with llama.cpp
Would love internvl-chat-v1-5 in a gguf format! https://internvl.opengvlab.com/
I second this
@ErfeiCui why did you close this as completed?
Any update on this? InternVL2-Llama3-76B on Ollama/llama.cpp would be amazing!
If someone gives me a tutorial I will write my own code to tranform this for pytorch to gguf for llama.cpp myself
It's more involved. you have to implement the model architecture and image preprocessing logic to llama.cpp which uses C++.