triton-inference-server topic

List triton-inference-server repositories

serving-codegen-gptj-triton

20
Stars
0
Forks
Watchers

Serving Example of CodeGen-350M-Mono-GPTJ on Triton Inference Server with Docker and Kubernetes

Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server -...

yolov8-triton

28
Stars
8
Forks
Watchers

Provides an ensemble model to deploy a YoloV8 ONNX model to Triton

GenerativeAIExamples

2.2k
Stars
428
Forks
36
Watchers

Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

recsys_pipeline

33
Stars
7
Forks
Watchers

Build Recommender System with PyTorch + Redis + Elasticsearch + Feast + Triton + Flask. Vector Recall, DeepFM Ranking and Web Application.

tritony

42
Stars
1
Forks
Watchers

Tiny configuration for Triton Inference Server

openai_trtllm

152
Stars
25
Forks
Watchers

OpenAI compatible API for TensorRT LLM triton backend

tensorrt-triton-magface

15
Stars
3
Forks
Watchers

Magface Triton Inferece Server Using Tensorrt