openchat
openchat copied to clipboard
llama_model_load: error loading model: create_tensor: tensor 'output.weight' not found
loaded openchat-3.5-0106-gemma-Q5_K_M.gguf ,then get error. it's newest version of llama.cpp
This is expected as Gemma uses tied embeddings, so the output.weight should be the same as input embeddings. Could you please open up an issue in llama.cpp repo?