llama.cpp
llama.cpp copied to clipboard
GGML_ASSERT: ggml.c:4014: false zsh: abort ./main -m ./models/65B/ggml-model-q4_0.bin -t 16 -n 256 --repeat_penalty 1.0
Not sure why this happens, I am on the latest commit and I am up-to-date on everything I did some tests and it seems like it breaks after 500~ tokens Is this a model limitation or can I fix this by increasing some value?
Please review and use our issue template to provide more details so we can better understand your problem and answer you.
Idk how to use it
idk how to get the check mark
now I'm running into another error, it segfaults with the regular script