24 Integrations with NVIDIA TensorRT

View a list of NVIDIA TensorRT integrations and software that integrates with NVIDIA TensorRT below. Compare the best NVIDIA TensorRT integrations as well as features, ratings, user reviews, and pricing of software that integrates with NVIDIA TensorRT. Here are the current NVIDIA TensorRT integrations in 2026:

  • 1
    Dataoorts GPU Cloud
    Dataoorts: Revolutionizing GPU Cloud Computing Dataoorts is a cutting-edge GPU cloud platform designed to meet the demands of the modern computational landscape. Launched in August 2024 after extensive beta testing, it offers revolutionary GPU virtualization technology, empowering researchers, developers, and businesses with unmatched flexibility, scalability, and performance. The Technology Behind Dataoorts At the core of Dataoorts lies its proprietary Dynamic Distributed Resource Allocation (DDRA) technology. This breakthrough allows real-time virtualization of GPU resources, ensuring optimal performance for diverse workloads. Whether you're training complex machine learning models, running high-performance simulations, or processing large datasets, Dataoorts delivers computational power with unparalleled efficiency.
    Leader badge
    Starting Price: $0.20/hour
  • 2
    TensorFlow

    TensorFlow

    TensorFlow

    An end-to-end open source machine learning platform. TensorFlow is an end-to-end open source platform for machine learning. It has a comprehensive, flexible ecosystem of tools, libraries and community resources that lets researchers push the state-of-the-art in ML and developers easily build and deploy ML powered applications. Build and train ML models easily using intuitive high-level APIs like Keras with eager execution, which makes for immediate model iteration and easy debugging. Easily train and deploy models in the cloud, on-prem, in the browser, or on-device no matter what language you use. A simple and flexible architecture to take new ideas from concept to code, to state-of-the-art models, and to publication faster. Build, deploy, and experiment easily with TensorFlow.
    Starting Price: Free
  • 3
    MATLAB

    MATLAB

    The MathWorks

    MATLAB® combines a desktop environment tuned for iterative analysis and design processes with a programming language that expresses matrix and array mathematics directly. It includes the Live Editor for creating scripts that combine code, output, and formatted text in an executable notebook. MATLAB toolboxes are professionally developed, rigorously tested, and fully documented. MATLAB apps let you see how different algorithms work with your data. Iterate until you’ve got the results you want, then automatically generate a MATLAB program to reproduce or automate your work. Scale your analyses to run on clusters, GPUs, and clouds with only minor code changes. There’s no need to rewrite your code or learn big data programming and out-of-memory techniques. Automatically convert MATLAB algorithms to C/C++, HDL, and CUDA code to run on your embedded processor or FPGA/ASIC. MATLAB works with Simulink to support Model-Based Design.
  • 4
    Python

    Python

    Python

    The core of extensible programming is defining functions. Python allows mandatory and optional arguments, keyword arguments, and even arbitrary argument lists. Whether you're new to programming or an experienced developer, it's easy to learn and use Python. Python can be easy to pick up whether you're a first-time programmer or you're experienced with other languages. The following pages are a useful first step to get on your way to writing programs with Python! The community hosts conferences and meetups to collaborate on code, and much more. Python's documentation will help you along the way, and the mailing lists will keep you in touch. The Python Package Index (PyPI) hosts thousands of third-party modules for Python. Both Python's standard library and the community-contributed modules allow for endless possibilities.
    Starting Price: Free
  • 5
    Rosepetal AI

    Rosepetal AI

    Rosepetal AI

    Rosepetal AI is an innovative technology company specializing in advanced artificial vision and deep-learning solutions designed specifically for industrial quality control. Our platform integrates dataset handling, automated labelling and training of adaptive neural networks, enabling real-time defect detection without requiring advanced technical expertise. This intuitive, no-code SaaS solution democratizes access to sophisticated AI, significantly enhancing efficiency, reducing waste, and driving operational excellence across multiple industries such as automotive, food processing, pharmaceuticals, plastics, and electronics. The unique strength of Rosepetal AI lies in its dynamic adaptability and scalability. Our system allows industrial companies to quickly deploy robust AI models directly onto their production lines, continuously adjusting to new product variations and emerging defects. This capability ensures consistent quality, minimizes downtime.
    Starting Price: €250
  • 6
    Hugging Face

    Hugging Face

    Hugging Face

    Hugging Face is a leading platform for AI and machine learning, offering a vast hub for models, datasets, and tools for natural language processing (NLP) and beyond. The platform supports a wide range of applications, from text, image, and audio to 3D data analysis. Hugging Face fosters collaboration among researchers, developers, and companies by providing open-source tools like Transformers, Diffusers, and Tokenizers. It enables users to build, share, and access pre-trained models, accelerating AI development for a variety of industries.
    Starting Price: $9 per month
  • 7
    NVIDIA Broadcast
    The NVIDIA Broadcast app transforms any room into a home studio. Take your live streams, voice chats, and video conference calls to the next level with AI-enhanced voice and video. Get a taste of what the Broadcast app can do for you using the incredible performance of NVIDIA® RTX GPUs. Loud keyboard typing, annoying microphone static, bothersome PC fans, nagging background reverb. Noise removal and room echo removal use AI to eliminate unwanted background noise at the touch of a button. Go live or attend a meeting remotely without worrying about finding a quiet place. Or mute players in loud environments, making them easier to understand. NVIDIA Broadcast makes it fast and easy. Customize your background without the need for expensive equipment or complex lighting setups. Personalize your webcam feed using powerful AI greenscreen effects perfect for work and play, wherever you are. The possibilities are endless.
    Starting Price: Free
  • 8
    RankLLM

    RankLLM

    Castorini

    RankLLM is a Python toolkit for reproducible information retrieval research using rerankers, with a focus on listwise reranking. It offers a suite of rerankers, pointwise models like MonoT5, pairwise models like DuoT5, and listwise models compatible with vLLM, SGLang, or TensorRT-LLM. Additionally, it supports RankGPT and RankGemini variants, which are proprietary listwise rerankers. It includes modules for retrieval, reranking, evaluation, and response analysis, facilitating end-to-end workflows. RankLLM integrates with Pyserini for retrieval and provides integrated evaluation for multi-stage pipelines. It also includes a module for detailed analysis of input prompts and LLM responses, addressing reliability concerns with LLM APIs and non-deterministic behavior in Mixture-of-Experts (MoE) models. The toolkit supports various backends, including SGLang and TensorRT-LLM, and is compatible with a wide range of LLMs.
    Starting Price: Free
  • 9
    RankGPT

    RankGPT

    Weiwei Sun

    RankGPT is a Python toolkit designed to explore the use of generative Large Language Models (LLMs) like ChatGPT and GPT-4 for relevance ranking in Information Retrieval (IR). It introduces methods such as instructional permutation generation and a sliding window strategy to enable LLMs to effectively rerank documents. It supports various LLMs, including GPT-3.5, GPT-4, Claude, Cohere, and Llama2 via LiteLLM. RankGPT provides modules for retrieval, reranking, evaluation, and response analysis, facilitating end-to-end workflows. It includes a module for detailed analysis of input prompts and LLM responses, addressing reliability concerns with LLM APIs and non-deterministic behavior in Mixture-of-Experts (MoE) models. The toolkit supports various backends, including SGLang and TensorRT-LLM, and is compatible with a wide range of LLMs. RankGPT's Model Zoo includes models like LiT5 and MonoT5, hosted on Hugging Face.
    Starting Price: Free
  • 10
    Kimi K2

    Kimi K2

    Moonshot AI

    Kimi K2 is a state-of-the-art open source large language model series built on a mixture-of-experts (MoE) architecture, featuring 1 trillion total parameters and 32 billion activated parameters for task-specific efficiency. Trained with the Muon optimizer on over 15.5 trillion tokens and stabilized by MuonClip’s attention-logit clamping, it delivers exceptional performance in frontier knowledge, reasoning, mathematics, coding, and general agentic workflows. Moonshot AI provides two variants, Kimi-K2-Base for research-level fine-tuning and Kimi-K2-Instruct pre-trained for immediate chat and tool-driven interactions, enabling both custom development and drop-in agentic capabilities. Benchmarks show it outperforms leading open source peers and rivals top proprietary models in coding tasks and complex task breakdowns, while its 128 K-token context length, tool-calling API compatibility, and support for industry-standard inference engines.
    Starting Price: Free
  • 11
    CUDA

    CUDA

    NVIDIA

    CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs. In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords. The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.
    Starting Price: Free
  • 12
    NVIDIA Merlin
    NVIDIA Merlin empowers data scientists, machine learning engineers, and researchers to build high-performing recommenders at scale. Merlin includes libraries, methods, and tools that streamline the building of recommenders by addressing common preprocessing, feature engineering, training, inference, and deploying to production challenges. Merlin components and capabilities are optimized to support the retrieval, filtering, scoring, and ordering of hundreds of terabytes of data, all accessible through easy-to-use APIs. With Merlin, better predictions, increased click-through rates, and faster deployment to production are within reach. NVIDIA Merlin, as part of NVIDIA AI, advances our commitment to supporting innovative practitioners doing their best work. As an end-to-end solution, NVIDIA Merlin components are designed to be interoperable within existing recommender workflows that utilize data science, and machine learning (ML).
  • 13
    NVIDIA virtual GPU
    NVIDIA virtual GPU (vGPU) software enables powerful GPU performance for workloads ranging from graphics-rich virtual workstations to data science and AI, enabling IT to leverage the management and security benefits of virtualization as well as the performance of NVIDIA GPUs required for modern workloads. Installed on a physical GPU in a cloud or enterprise data center server, NVIDIA vGPU software creates virtual GPUs that can be shared across multiple virtual machines, and accessed by any device, anywhere. Deliver performance virtually indistinguishable from a bare metal environment. Leverage common data center management tools such as live migration. Provision GPU resources with fractional or multi-GPU virtual machine (VM) instances. Responsive to changing business requirements and remote teams.
  • 14
    NVIDIA NIM
    Explore the latest optimized AI models, connect AI agents to data with NVIDIA NeMo, and deploy anywhere with NVIDIA NIM microservices. NVIDIA NIM is a set of easy-to-use inference microservices that facilitate the deployment of foundation models across any cloud or data center, ensuring data security and streamlined AI integration. Additionally, NVIDIA AI provides access to the Deep Learning Institute (DLI), offering technical training to gain in-demand skills, hands-on experience, and expert knowledge in AI, data science, and accelerated computing. AI models generate responses and outputs based on complex algorithms and machine learning techniques, and those responses or outputs may be inaccurate, harmful, biased, or indecent. By testing this model, you assume the risk of any harm caused by any response or output of the model. Please do not upload any confidential information or personal data unless expressly permitted. Your use is logged for security purposes.
  • 15
    NVIDIA Jetson
    NVIDIA's Jetson platform is a leading solution for embedded AI computing, utilized by professional developers to create breakthrough AI products across various industries, as well as by students and enthusiasts for hands-on AI learning and innovative projects. The platform comprises small, power-efficient production modules and developer kits, offering a comprehensive AI software stack for high-performance acceleration. This enables the deployment of generative AI at the edge, supporting applications like NVIDIA Metropolis and the Isaac platform. The Jetson family includes a range of modules tailored to different performance and power efficiency needs, such as the Jetson Nano, Jetson TX2, Jetson Xavier NX, and the Jetson Orin series. Each module is designed to meet specific AI computing requirements, from entry-level projects to advanced robotics and industrial applications.
  • 16
    PyTorch

    PyTorch

    PyTorch

    Transition seamlessly between eager and graph modes with TorchScript, and accelerate the path to production with TorchServe. Scalable distributed training and performance optimization in research and production is enabled by the torch-distributed backend. A rich ecosystem of tools and libraries extends PyTorch and supports development in computer vision, NLP and more. PyTorch is well supported on major cloud platforms, providing frictionless development and easy scaling. Select your preferences and run the install command. Stable represents the most currently tested and supported version of PyTorch. This should be suitable for many users. Preview is available if you want the latest, not fully tested and supported, 1.10 builds that are generated nightly. Please ensure that you have met the prerequisites (e.g., numpy), depending on your package manager. Anaconda is our recommended package manager since it installs all dependencies.
  • 17
    NVIDIA Riva Studio
    Use the browser with in-app prompts and a recording tool. A predefined set of phonetically balanced sentences is available to create a 30-minute dataset for training a TTS model to learn your unique voice. Make the model sound like you by choosing the range that best suits the pitch of your voice. The recommended typical voice pitch range setting for a human voice is already provided, along with a preprogrammed best recipe to customize the TTS model for your voice. Generate an API to integrate a customized TTS model into your application. Download a deployable package with a helm chart to run on any cloud or on-premises Kubernetes cluster. Then, automatically host your voice microservice with NVIDIA, or set it up with just one-line of code. Set up, customize, and deploy the Riva TTS model with intuitive no-code, end-to-end GUI workflows and no infrastructure configuration.
  • 18
    NVIDIA AI Enterprise
    The software layer of the NVIDIA AI platform, NVIDIA AI Enterprise accelerates the data science pipeline and streamlines development and deployment of production AI including generative AI, computer vision, speech AI and more. With over 50 frameworks, pretrained models and development tools, NVIDIA AI Enterprise is designed to accelerate enterprises to the leading edge of AI, while also simplifying AI to make it accessible to every enterprise. The adoption of artificial intelligence and machine learning has gone mainstream, and is core to nearly every company’s competitive strategy. One of the toughest challenges for enterprises is the struggle with siloed infrastructure across the cloud and on-premises data centers. AI requires their environments to be managed as a common platform, instead of islands of compute.
  • 19
    NVIDIA Clara
    Clara’s domain-specific tools, AI pre-trained models, and accelerated applications are enabling AI breakthroughs in numerous fields, including medical devices, imaging, drug discovery, and genomics. Explore the end-to-end pipeline of medical device development and deployment with the Holoscan platform. Build containerized AI apps with the Holoscan SDK and MONAI, and streamline deployment in next-generation AI devices with the NVIDIA IGX developer kits. The NVIDIA Holoscan SDK includes healthcare-specific acceleration libraries, pre-trained AI models, and reference applications for computational medical devices.
  • 20
    NVIDIA Morpheus
    NVIDIA Morpheus is a GPU-accelerated, end-to-end AI framework that enables developers to create optimized applications for filtering, processing, and classifying large volumes of streaming cybersecurity data. Morpheus incorporates AI to reduce the time and cost associated with identifying, capturing, and acting on threats, bringing a new level of security to the data center, cloud, and edge. Morpheus also extends human analysts’ capabilities with generative AI by automating real-time analysis and responses, producing synthetic data to train AI models that identify risks accurately and run what-if scenarios. Morpheus is available as open-source software on GitHub for developers interested in using the latest pre-release features and who want to build from source. Get unlimited usage on all clouds, access to NVIDIA AI experts, and long-term support for production deployments with a purchase of NVIDIA AI Enterprise.
  • 21
    LaunchX

    LaunchX

    Nota AI

    Optimized AI is ready to launch on-device and allows you to deploy your AI models on actual devices. With LaunchX automation, you can simplify conversion and effortlessly measure performance on target devices. Customize the AI platform to meet your hardware specifications. Enable seamless AI model deployment with a tailored software stack. Nota’s AI technology empowers intelligent transportation systems, facial recognition, and security and surveillance. The company’s solutions include a driver monitoring system, driver authentication, and smart access control system. Nota‘s current projects cover a wide range of industries including construction, mobility, security, smart home, and healthcare. Nota’s partnership with top-tier global market leaders including Nvidia, Intel, and ARM has helped accelerate its entry into the global market.
  • 22
    NVIDIA DeepStream SDK
    NVIDIA's DeepStream SDK is a comprehensive streaming analytics toolkit based on GStreamer, designed for AI-based multi-sensor processing, including video, audio, and image understanding. It enables developers to create stream-processing pipelines that incorporate neural networks and complex tasks like tracking, video encoding/decoding, and rendering, facilitating real-time analytics on various data types. DeepStream is integral to NVIDIA Metropolis, a platform for building end-to-end services that transform pixel and sensor data into actionable insights. The SDK offers a powerful and flexible environment suitable for a wide range of industries, supporting multiple programming options such as C/C++, Python, and Graph Composer's intuitive UI. It allows for real-time insights by understanding rich, multi-modal sensor data at the edge and supports managed AI services through deployment in cloud-native containers orchestrated with Kubernetes.
  • 23
    Ultralytics

    Ultralytics

    Ultralytics

    Ultralytics offers a full-stack vision-AI platform built around its flagship YOLO model suite that enables teams to train, validate, and deploy computer-vision models with minimal friction. The platform allows you to drag and drop datasets, select from pre-built templates or fine-tune custom models, then export to a wide variety of formats for cloud, edge or mobile deployment. With support for tasks including object detection, instance segmentation, image classification, pose estimation and oriented bounding-box detection, Ultralytics’ models deliver high accuracy and efficiency and are optimized for both embedded devices and large-scale inference. The product also includes Ultralytics HUB, a web-based tool where users can upload their images/videos, train models online, preview results (even on a phone), collaborate with team members, and deploy via an inference API.
  • 24
    NVIDIA DRIVE
    Software is what turns a vehicle into an intelligent machine. The NVIDIA DRIVE™ Software stack is open, empowering developers to efficiently build and deploy a variety of state-of-the-art AV applications, including perception, localization and mapping, planning and control, driver monitoring, and natural language processing. The foundation of the DRIVE Software stack, DRIVE OS is the first safe operating system for accelerated computing. It includes NvMedia for sensor input processing, NVIDIA CUDA® libraries for efficient parallel computing implementations, NVIDIA TensorRT™ for real-time AI inference, and other developer tools and modules to access hardware engines. The NVIDIA DriveWorks® SDK provides middleware functions on top of DRIVE OS that are fundamental to autonomous vehicle development. These consist of the sensor abstraction layer (SAL) and sensor plugins, data recorder, vehicle I/O support, and a deep neural network (DNN) framework.
  • Previous
  • You're on page 1
  • Next