← Registry

triton

Community

NVIDIA Triton Inference Server for deploying AI models at scale. Supports multiple frameworks (ONNX, TensorRT, PyTorch, TensorFlow), model ensembles, dynamic batching, model versioning, and GPU/CPU inference with high throughput and low latency.

Install

skillpm install triton

Format score

100/100

Spec

v1.0

Installs

0

Published

April 1, 2026