Wenlong Shi
Wenlong Shi
llama.cpp is widely recognized for deploying LLM, including Gemma. Have there been any performance comparisons between gemma.cpp and llama.cpp?
This project has not been updated for almost three months. Has the project been deprecated? Will the planned tasks in the roadmap be implemented?
I encountered a test failure after building and running the tests. Here are the details: - **GPU**: RTX 4090 - **Repo branch**: v1.4.0 - **Operating System**: Ubuntu 22.04.3 - **CUDA...