Results 2 issues of Anirudh JM

I got it to works just like instruction, I'm using CUDA 12.3: `set CMAKE_ARGS="-DLLAMA_CUBLAS=on" && set FORCE_CMAKE=1 && pip install --no-cache-dir llama-cpp-python==0.2.90 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu123` _Originally posted by @BinhPQ2 in https://github.com/abetlen/llama-cpp-python/issues/576#issuecomment-2379861701_

# Prerequisites Please answer the following questions for yourself before submitting an issue. - [ Yes] I am running the latest code. Development is very rapid so there are no...