llama-cpp-python
llama-cpp-python copied to clipboard
unknown model architecture: 'gemma-embedding'
I am running llama-cpp-python Version: 0.3.16
Trying to load the recently released model embeddinggemma-300M I get the following error message:
llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'gemma-embedding'
Support for this model architecture has been added to llama.cpp in this build: https://github.com/ggml-org/llama.cpp/releases/tag/b6384
Could you please align llama-cpp-python to reflect this addition ?
I tried to just dump in the latest llama.cpp under vendor and build, resulted in this error when trying to load the model:
.local/lib/python3.13/site-packages/llama_cpp/llama_cpp.py", line 1408, in
So... nothing being done on this? I have same issue as OP.
The error is because support for gemma embedding was added to the original llama.cpp library recently, and the llama-cpp-python binding is not up-to-date with the original one.
Blessings, any update?