Skip to content
Change the repository type filter

All

    Repositories list

    • server

      Public
      The Triton Inference Server provides an optimized cloud and edge inferencing solution.
      Python
      BSD 3-Clause "New" or "Revised" License
      1.5k8.9k66273Updated Mar 22, 2025Mar 22, 2025
    • C++
      BSD 3-Clause "New" or "Revised" License
      1351612Updated Mar 21, 2025Mar 21, 2025
    • Python
      BSD 3-Clause "New" or "Revised" License
      2423706Updated Mar 21, 2025Mar 21, 2025
    • core

      Public
      The core library and APIs implementing the Triton Inference Server.
      C++
      BSD 3-Clause "New" or "Revised" License
      106122018Updated Mar 21, 2025Mar 21, 2025
    • tutorials

      Public
      This repository contains tutorials and examples for Triton Inference Server
      Python
      BSD 3-Clause "New" or "Revised" License
      112671815Updated Mar 21, 2025Mar 21, 2025
    • Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
      C++
      BSD 3-Clause "New" or "Revised" License
      161595011Updated Mar 21, 2025Mar 21, 2025
    • common

      Public
      Common source, scripts and utilities shared across all Triton repositories.
      C++
      BSD 3-Clause "New" or "Revised" License
      746906Updated Mar 21, 2025Mar 21, 2025
    • The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.
      C++
      MIT License
      32132217Updated Mar 21, 2025Mar 21, 2025
    • client

      Public
      Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.
      Python
      BSD 3-Clause "New" or "Revised" License
      2376124025Updated Mar 20, 2025Mar 20, 2025
    • Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inference Server.
      Python
      46022Updated Mar 19, 2025Mar 19, 2025
    • The Triton TensorRT-LLM Backend
      Python
      Apache License 2.0
      11980730723Updated Mar 18, 2025Mar 18, 2025
    • backend

      Public
      Common source, scripts and utilities for creating Triton backends.
      C++
      BSD 3-Clause "New" or "Revised" License
      9531003Updated Mar 17, 2025Mar 17, 2025
    • The Triton backend for the ONNX Runtime.
      C++
      BSD 3-Clause "New" or "Revised" License
      61140742Updated Mar 14, 2025Mar 14, 2025
    • FIL backend for the Triton Inference Server
      Jupyter Notebook
      Apache License 2.0
      3676512Updated Mar 13, 2025Mar 13, 2025
    • Third-party source packages that are modified for use in Triton.
      C
      BSD 3-Clause "New" or "Revised" License
      58704Updated Mar 12, 2025Mar 12, 2025
    • The Triton backend for TensorRT.
      C++
      BSD 3-Clause "New" or "Revised" License
      327001Updated Mar 12, 2025Mar 12, 2025
    • The Triton backend for TensorFlow.
      C++
      BSD 3-Clause "New" or "Revised" License
      225102Updated Mar 12, 2025Mar 12, 2025
    • Simple Triton backend used for testing.
      C++
      BSD 3-Clause "New" or "Revised" License
      5200Updated Mar 12, 2025Mar 12, 2025
    • An example Triton backend that demonstrates sending zero, one, or multiple responses for each request.
      C++
      BSD 3-Clause "New" or "Revised" License
      7500Updated Mar 12, 2025Mar 12, 2025
    • TRITONCACHE implementation of a Redis cache
      C++
      BSD 3-Clause "New" or "Revised" License
      41320Updated Mar 12, 2025Mar 12, 2025
    • The Triton backend for the PyTorch TorchScript models.
      C++
      BSD 3-Clause "New" or "Revised" License
      4814404Updated Mar 12, 2025Mar 12, 2025
    • OpenVINO backend for Triton.
      C++
      BSD 3-Clause "New" or "Revised" License
      173163Updated Mar 12, 2025Mar 12, 2025
    • Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
      Python
      Apache License 2.0
      78462266Updated Mar 12, 2025Mar 12, 2025
    • Implementation of a local in-memory cache for Triton Inference Server's TRITONCACHE API
      C++
      BSD 3-Clause "New" or "Revised" License
      1510Updated Mar 12, 2025Mar 12, 2025
    • Example Triton backend that demonstrates most of the Triton Backend API.
      C++
      BSD 3-Clause "New" or "Revised" License
      12700Updated Mar 12, 2025Mar 12, 2025
    • C++
      101805Updated Mar 12, 2025Mar 12, 2025
    • The Triton repository agent that verifies model checksums.
      C++
      BSD 3-Clause "New" or "Revised" License
      71100Updated Mar 12, 2025Mar 12, 2025
    • triton_distributed

      Public archive
      Rust
      Apache License 2.0
      14493637Updated Mar 7, 2025Mar 7, 2025
    • .github

      Public
      Community health files for NVIDIA Triton
      1100Updated Feb 27, 2025Feb 27, 2025
    • pytriton

      Public
      PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.
      Python
      Apache License 2.0
      53781110Updated Feb 12, 2025Feb 12, 2025