Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate into the JIT runtime seamlessly. After compilation using the optimized graph should feel no different than running a TorchScript module. You also have access to TensorRT’s suite of configurations at compile time, so you are able to specify operating precision (FP32/FP16/INT8) and other settings for your module.

Features

  • Build a docker container for Torch-TensorRT
  • NVIDIA NGC Container
  • Requires Libtorch 1.12.0 (built with CUDA 11.3)
  • Build using cuDNN & TensorRT tarball distributions
  • Test using Python backend
  • You have access to TensorRT's suite of configurations at compile time

Project Samples

Project Activity

See All Activity >

License

BSD License

Follow Torch-TensorRT

Torch-TensorRT Web Site

Other Useful Business Software
MongoDB Atlas runs apps anywhere Icon
MongoDB Atlas runs apps anywhere

Deploy in 115+ regions with the modern database for every enterprise.

MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
Start Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Torch-TensorRT!

Additional Project Details

Programming Language

C++

Related Categories

C++ Machine Learning Software, C++ Deep Learning Frameworks

Registered

2022-08-12