booster
booster copied to clipboard
Booster - open accelerator for LLM models. Better inference and debugging for AI hackers
Large Hadron Collider is the world's most powerful particle accelerator.
Large Model Collider aims to be an simple and mighty LLM inference accelerator both for those who needs to scale GPTs within production environment or just experiment with models on its own.
TLDR
☝ Latest build v1.2 for Mac Apple Silicon
M1/M2/M3 CPUs and integrated GPUs supported, just place ggml-metal.metal into the same folder:
collider-v1.2.0-mac + ggml-metal.metal
☝ Latest build v1.2 for Linux / CUDA
It was buit on Ubuntu 22.02 with CUDA 12.1 drivers. Should work on Intel / AMD Linux platforms:
Superpowers
- Built with performance and scaling in mind thanks Golang and C++
- No more problems with Python dependencies and broken compatibility
- Most of modern CPUs are supported: any Intel/AMD x64 platofrms, server and Mac ARM64
- GPUs supported as well: Nvidia CUDA, Apple Metal, OpenCL cards
- Split really big models between a number of GPU (warp LLaMA 70B with 2x RTX 3090)
- Not bad performance on shy CPU machines, fast as hell inference on monsters with beefy GPUs
- Both regular FP16/FP32 models and their quantised versions are supported - 4-bit really rocks!
- Popular LLM architectures already there: LLaMA, Starcoder, Baichuan, Mistral, etc...
- Special bonus: proprietary Janus Sampling for code generation and non English languages
Motivation
Within first month of llama.go development I was literally shocked of how original ggml.cpp project made it very clear - there are no limits for talented people on bringing mind-blowing features and moving to AI future.
So I've decided to start a new project where best-in-class C++ / CUDA core will be embedded into mighty Golang server ready for robust and performant inference at large scale within real production environments.
V0 Roadmap - Fall'23
- [x] Draft implementation with CGO llama.cpp backend
- [x] Simple REST API to allow text generation
- [x] Inference with Apple Silicon GPU using Metal framework
- [x] Parallel inference both with CPU and GPU
- [x] Support both AMD64 and ARM64 platforms
- [x] CUDA support and fast inference with Nvidia cards
- [x] Retain dialog history by Session ID parameter
- [x] Support moderm GGUF V3 model format
- [x] Inference for most popular LLM architectures
- [x] Janus Sampling for better non-English text generation
V1 Roadmap - Winter'23
- [x] Rebrand project: LLaMAZoo => Large Model Collider
- [x] Is it 2023, 30th of November? First birthday of ChatGPT! Celebrate ...
- [x] ... then release Collider V1 after half a year of honing it :)
- [ ] Freeze JSON / YAML config format for Native API
- [ ] Release OpenAI API compatible endpoints
- [ ] Perplexity computation [ useful for benchmarking ]
- [ ] Support LLaVA multi-modal models inference
V2 Roadmap - Spring'24
- [ ] Full Windows support
- [ ] Prebuilt binaries for all platforms
- [ ] Better test coverage
How to build on Mac?
Collider was (and still) developed on Mac with Apple Silicon M1 processor, so it's really easy peasy:
make mac
How to compile for CUDA on Ubuntu?
Follow step 1 and step 2, then just make!
Ubuntu Step 1: Install C++ and Golang compilers, as well some developer libraries
sudo apt update -y && sudo apt upgrade -y && \
apt install -y git git-lfs make build-essential && \
wget https://golang.org/dl/go1.21.5.linux-amd64.tar.gz && \
tar -xf go1.21.5.linux-amd64.tar.gz -C /usr/local && \
rm go1.21.5.linux-amd64.tar.gz && \
echo 'export PATH="${PATH}:/usr/local/go/bin"' >> ~/.bashrc && source ~/.bashrc
Ubuntu Step 2: Install Nvidia drivers and CUDA Toolkit 12.2 with NVCC
wget https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/cuda-ubuntu2204.pin && \
sudo mv cuda-ubuntu2204.pin /etc/apt/preferences.d/cuda-repository-pin-600 && \
sudo apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub && \
sudo add-apt-repository "deb http://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /" && \
sudo apt update -y && \
sudo apt install -y cuda-toolkit-12-2
Now you are ready to rock
make cuda
How to Run?
You shold go through steps below:
- Build the server from sources [ pure CPU inference as example ]
make clean && make mac
- Download the model [ like Mistral 7B quantized to GGUF Q4KM format as an example ]
wget https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q4_K_M.gguf
- Create configuration file and place it to the same directory [ see config.sample.yaml ]
id: "collider"
host: localhost
port: 8080
log: collider.log
deadline: 180
debug: full
swap:
pods:
-
model: default
threads: 6
gpus: [ 0 ]
batchsize: 512
models:
-
id: default
name: Mistral
path: mistral-7b-instruct-v0.1.Q4_K_M.gguf
locale: en_US
preamble: "You are a virtual assistant. Please answer the question."
prefix: "\nUSER: "
suffix: "\nASSISTANT:"
contextsize: 2048
predict: 1024
temperature: 0.1
top_k: 8
top_p: 0.96
repetition_penalty: 1.1
- When all is done, start the server with debug enabled to be sure it working
./collider --server --debug
- Now POST JSON with unique ID and your question to
localhost:8080/jobs
{
"id": "5fb8ebd0-e0c9-4759-8f7d-35590f6c9fc6",
"prompt": "Who are you?"
}
- See instructions within
collider.service
file on how to create daemond service out of this API server.