Audience

Developers and companies searching for an inference server solution to improve AI production

About NVIDIA Triton Inference Server

NVIDIA Triton™ inference server delivers fast and scalable AI in production. Open-source inference serving software, Triton inference server streamlines AI inference by enabling teams deploy trained AI models from any framework (TensorFlow, NVIDIA TensorRT®, PyTorch, ONNX, XGBoost, Python, custom and more on any GPU- or CPU-based infrastructure (cloud, data center, or edge). Triton runs models concurrently on GPUs to maximize throughput and utilization, supports x86 and ARM CPU-based inferencing, and offers features like dynamic batching, model analyzer, model ensemble, and audio streaming. Triton helps developers deliver high-performance inference aTriton integrates with Kubernetes for orchestration and scaling, exports Prometheus metrics for monitoring, supports live model updates, and can be used in all major public cloud machine learning (ML) and managed Kubernetes platforms. Triton helps standardize model deployment in production.

Pricing

Starting Price:
Free
Free Version:
Free Version available.

Integrations

Ratings/Reviews

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Company Information

NVIDIA
United States
developer.nvidia.com/nvidia-triton-inference-server

Videos and Screen Captures

You Might Also Like
Our Free Plans just got better! | Auth0 by Okta Icon
Our Free Plans just got better! | Auth0 by Okta

With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your secuirty. Auth0 now, thank yourself later.
Try free now

Product Details

Platforms Supported
Windows
Mac
Linux
Training
Documentation
In Person
Videos
Support
Phone Support
Online

NVIDIA Triton Inference Server Frequently Asked Questions

Q: What kinds of users and organization types does NVIDIA Triton Inference Server work with?
Q: What languages does NVIDIA Triton Inference Server support in their product?
Q: What kind of support options does NVIDIA Triton Inference Server offer?
Q: What other applications or services does NVIDIA Triton Inference Server integrate with?
Q: What type of training does NVIDIA Triton Inference Server provide?
Q: How much does NVIDIA Triton Inference Server cost?

NVIDIA Triton Inference Server Product Features

Artificial Intelligence

For Healthcare
Multi-Language
Chatbot
For Sales
Rules-Based Automation
Machine Learning
Natural Language Processing
Predictive Analytics
Virtual Personal Assistant (VPA)
Process/Workflow Automation
For eCommerce
Image Recognition

Machine Learning

Deep Learning
ML Algorithm Library
Model Training
Natural Language Processing (NLP)
Predictive Modeling
Templates
Visualization
Statistical / Mathematical Tools

NVIDIA Triton Inference Server Additional Categories