BitNet icon indicating copy to clipboard operation
BitNet copied to clipboard

Official inference framework for 1-bit LLMs

Results 67 BitNet issues
Sort by recently updated
recently updated
newest added

run : python setup_env.py -md /home/disk1/Llama3-8B-1.58-100B-tokens -q i2_s found: FileNotFoundError: [Errno 2] No such file or directory: './build/bin/llama-quantize'

can we generate the `llama-server` too? is not listed after the build \/ ![Image](https://github.com/user-attachments/assets/5c51c8a4-be0c-45f2-bd98-95e6fd02e88d)

I manually downloaded the model and set the model with the command "python setup_env.py -md .\models\Llama3-8B-1.58-100B-tokens -q i2_s" in Windows 11 OS. The result shows: "ERROR:root:Error occurred while running command:...

When quantizing the (input) activations to the bit-linear layer, `NaN`s may occur due to division by zero. This is a consequence of the formula in the original paper: $Quant(x) =...

Basically same as #84 but for TL1 codegen.

I have just discovered bitnet. Before this, I was using models in gguf format. But they were below my expectations. Now I want to run the models I trained on...

Hi BitNet team, Thanks for releasing this great work! I'm a phd student at UC Irvine. Our research is primarily focusing on LLM efficient deployment on mobile devices. I wish...