llama.cpp
llama.cpp copied to clipboard
gguf_convert_endian.py: implement byteswapping for q4_k and q6_k
With these changes llama3.2 model could be converted to big endian.