BitNet icon indicating copy to clipboard operation
BitNet copied to clipboard

Official inference framework for 1-bit LLMs

Results 67 BitNet issues
Sort by recently updated
recently updated
newest added

I am quite a newbie to LLM's, but was running into the following issue: Command: `python setup_env.py --hf-repo HF1BitLLM/Llama3-8B-1.58-100B-tokens -q i2_s --quant-embd` Output: `ERROR:root:Error occurred while running command: Command '['/home/pis7/miniconda3/envs/bitnet-cpp/bin/python',...

In file included from /root/BitNet/3rdparty/llama.cpp/ggml/src/./ggml-quants.h:4, from /root/BitNet/src/ggml-bitnet-lut.cpp:9: /root/BitNet/3rdparty/llama.cpp/ggml/src/./ggml-common.h:154:16: warning: ISO C++ prohibits anonymous structs [-Wpedantic] 154 | struct { | ^ /root/BitNet/3rdparty/llama.cpp/ggml/src/./ggml-common.h:175:16: warning: ISO C++ prohibits anonymous structs [-Wpedantic] 175...

## Description Error when running the llama-bench tool on a dummy model, as specified in readme. Failed assertion in ggml.c relating to tile number for parallel processing. ## Steps to...

I followed [Basic Usage section](https://github.com/microsoft/BitNet?tab=readme-ov-file#basic-usage) in documentation, and it appears that when sending very basic prompts for completion, the Model returns irrelevant completions/answers in the good scenario, and in worst...

Add a command to pull git submodule `3rdparty` repository before running pip install, so the pip install command will run with latest updates from submodule branch ( for cases in...

- python>=3.9 - cmake>=3.22 - clang>=18 on windows use the msvc-clang (>=17) fix https://github.com/microsoft/BitNet/issues/34

Hi all, I have tested the ggml-model-tl1.gguf (Llama3-8B-1.58-100B-tokens) model on iOS (iPhone 15 Pro) using the llamaSwiftUI sample project. However, I encountered an EXC_BAD_ACCESS error at ggml_vec_dot_f32(). Can you confirm...

Can you support reordering and vector embedding models?

Functionality similar to LLaMA.cpp HTTP Server. ./llama-server

I've been working on securing the user input, escaping invalid characters, however I've encountered a few prompts which cause the llama-cli to abruptly halt: ``` .\llama-cli.exe --model "..\..\..\models\Llama3-8B-1.58-100B-tokens\ggml-model-i2_s.gguf" --prompt "£"...