Triton Inference Server
Triton Inference Server
server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
client
Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.
model_analyzer
Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
python_backend
Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
onnxruntime_backend
The Triton backend for the ONNX Runtime.
dali_backend
The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.
model_navigator
Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.
backend
Common source, scripts and utilities for creating Triton backends.
common
Common source, scripts and utilities shared across all Triton repositories.