Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
tritonserver2.58.0-igpu.tar | 2025-05-31 | 1.6 GB | |
v2.58.0_ubuntu2404.clients.tar.gz | 2025-05-31 | 180.7 MB | |
README.md | 2025-05-28 | 10.3 kB | |
Release 2.58.0 corresponding to NGC container 25.05 source code.tar.gz | 2025-05-28 | 9.8 MB | |
Release 2.58.0 corresponding to NGC container 25.05 source code.zip | 2025-05-28 | 11.5 MB | |
Totals: 5 Items | 1.8 GB | 0 |
Triton Inference Server
The Triton Inference Server provides a cloud inferencing solution optimized for both CPUs and GPUs. The server provides an inference service via an HTTP or GRPC endpoint, allowing remote clients to request inferencing for any model being managed by the server. For edge deployments, Triton Server is also available as a shared library with an API that allows the full functionality of the server to be included directly in an application.