Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate into the JIT runtime seamlessly. After compilation using the optimized graph should feel no different than running a TorchScript module. You also have access to TensorRT’s suite of configurations at compile time, so you are able to specify operating precision (FP32/FP16/INT8) and other settings for your module.

Features

  • Build a docker container for Torch-TensorRT
  • NVIDIA NGC Container
  • Requires Libtorch 1.12.0 (built with CUDA 11.3)
  • Build using cuDNN & TensorRT tarball distributions
  • Test using Python backend
  • You have access to TensorRT's suite of configurations at compile time

Project Samples

Project Activity

See All Activity >

License

BSD License

Follow Torch-TensorRT

Torch-TensorRT Web Site

Other Useful Business Software
Earn up to 16% annual interest with Nexo. Icon
Earn up to 16% annual interest with Nexo.

Access competitive interest rates on your digital assets.

Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
Get started with Nexo.
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Torch-TensorRT!

Additional Project Details

Programming Language

C++

Related Categories

C++ Machine Learning Software, C++ Deep Learning Frameworks

Registered

2022-08-12