ggml topic
llm
[Unmaintained, see README] An ecosystem of Rust libraries for working with large language models
mpt-30B-inference
Run inference on MPT-30B using CPU
inference
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready...
ggml-js
JavaScript bindings for the ggml-js library
minigpt4.cpp
Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)
clip.cpp
CLIP inference in plain C/C++ with no extra dependencies
gpu_poor
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
ialacol
🪶 Lightweight OpenAI drop-in replacement for Kubernetes
LLaMA-Cult-and-More
Large Language Models for All, 🦙 Cult and More, Stay in touch !
py.gpt.prompt
PyGPTPrompt: A CLI tool that manages context windows for AI models, facilitating user interaction and data ingestion for optimized long-term memory and task automation.