Juan Fran

Results 1 issues of Juan Fran

Hello Since latest changes in model quantization in llama.cpp, I am not able to load any model in GPU memory. When trying to load an older quantization model like vicuna-7b-1.1.ggmlv3.q4_0.bin...

bug