Model conversion from HF to GGUF crashes due to lack of memory
This was for the 8B param model in the instructions - quickly ran through the 32GB RAM on my Linux PC. Is there someplace that pre-converted GGUF versions of these models might reside so that this conversion wouldn't need to be done?
https://huggingface.co/brunopio/Llama3-8B-1.58-100B-tokens-GGUF
Same issue here, with my M2 Macbook Air with 16Gb RAM worked. Linux 32Gb get killed by the SO 🤔
On Linux, I solved by using a 32Gb RAM machine + 10Gb Swap.
Please try with our latest release gguf file https://huggingface.co/microsoft/bitnet-b1.58-2B-4T-gguf. thanks.