llm-inference topic
sagify
LLMs and Machine Learning done easily
GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
runbooks
Finetune LLMs on K8s by using Runbooks
lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
llm-api
Run any Large Language Model behind a unified API
local-llm-function-calling
A tool for generating function arguments and choosing what function to call with local LLMs
nos
⚡️ A fast and flexible PyTorch inference server that runs locally, on any cloud or AI HW.
LLM-Minutes-of-Meeting
🎤📄 An innovative tool that transforms audio or video files into text transcripts and generates concise meeting minutes. Stay organized and efficient in your meetings, and get ready for Phase 2 where...
neural-speed
An innovative library for efficient LLM inference via low-bit quantization
PalmHill.BlazorChat
PalmHill.BlazorChat is a chat application and API built with Blazor WebAssembly, SignalR, and WebAPI, featuring real-time LLM conversations, markdown support, customizable settings, and a responsive d...