BentoML
BentoML
BentoML
The easiest way to serve AI apps and models - Build reliable Inference APIs, LLM apps, Multi-model chains, RAG service, and much more!
Yatai
Model Deployment at Scale on Kubernetes 🦄️
gallery
BentoML Example Projects 🎨
bentoctl
Fast model deployment on any cloud 🚀
stable-diffusion-server
Deploy Your Own Stable Diffusion Service
OpenLLM
Run any open-source LLMs, such as Llama 3.1, Gemma, as OpenAI compatible API endpoint in the cloud.
transformers-nlp-service
Online Inference API for NLP Transformer models - summarization, text classification, sentiment analysis and more
OneDiffusion
OneDiffusion: Run any Stable Diffusion models and fine-tuned weights with ease
OCR-as-a-Service
Turn any OCR models into online inference API endpoint 🚀 🌖
CLIP-API-service
CLIP as a service - Embed image and sentences, object recognition, visual reasoning, image classification and reverse image search