Juan Fran
Results
1
issues of
Juan Fran
Hello Since latest changes in model quantization in llama.cpp, I am not able to load any model in GPU memory. When trying to load an older quantization model like vicuna-7b-1.1.ggmlv3.q4_0.bin...
bug