Cui Junbo
Cui Junbo
Thanks a lot for the feedback, we also found a difference between the llamacpp and int4 versions. We are trying to find the problem. @naifmeh
> Hi naifmeh, to fix the above code you can do this on the file `minicpmv.cpp` which is located under `examples/minicpmv` > > There what you can do is change...
@naifmeh now we have solved this problem, please try it, looking forward to your feedback!
@tc-mb Have a look.
MiniCPM-Llama3-V 2.5 can run with llama.cpp now! See our fork of [llama.cpp](https://github.com/OpenBMB/llama.cpp/blob/minicpm-v2.5/examples/minicpmv/README.md) for more detail. and here is our model in gguf format. https://huggingface.co/openbmb/MiniCPM-Llama3-V-2_5-gguf @duanshuaimin @leeaction @tyzero @hexf00
> 可否出一個 F16的 gguf格式,Q4的量化 沒有吃滿顯存。 > > 我們嘗試按照readme轉換一了下,發現效果也不是很好,還請指正。 > > [readme](https://github.com/OpenBMB/llama.cpp/blob/minicpm-v2.5/examples/minicpmv/README.md) > >  > >  It looks like you're using ollama instead of the direct llama.cpp? The current...
@seasoncool https://github.com/OpenBMB/llama.cpp Please use our fork, the offical llama.cpp has not merge our PR
https://github.com/OpenBMB/ollama/tree/minicpm-v2.5/examples/minicpm-v2.5 @yuanjie-ai
We notice some reported issues from compromising MiniCPM-Llama3-V 2.5's adaptive visual encoding with Ollama & Llama.cpp's vanilla fixed encoding implementation. We are reimplementing this part for Ollama & Llama.cpp to...
This issue does not provide a reproducible context and requires more information to help resolve it. If you still need assistance, please provide your code environment and running code to...