Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate into the JIT runtime seamlessly. After compilation using the optimized graph should feel no different than running a TorchScript module. You also have access to TensorRT’s suite of configurations at compile time, so you are able to specify operating precision (FP32/FP16/INT8) and other settings for your module.

Features

  • Build a docker container for Torch-TensorRT
  • NVIDIA NGC Container
  • Requires Libtorch 1.12.0 (built with CUDA 11.3)
  • Build using cuDNN & TensorRT tarball distributions
  • Test using Python backend
  • You have access to TensorRT's suite of configurations at compile time

Project Samples

Project Activity

See All Activity >

License

BSD License

Follow Torch-TensorRT

Torch-TensorRT Web Site

Other Useful Business Software
The Most Powerful Software Platform for EHSQ and ESG Management Icon
The Most Powerful Software Platform for EHSQ and ESG Management

Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
Learn More
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Torch-TensorRT!

Additional Project Details

Programming Language

C++

Related Categories

C++ Machine Learning Software, C++ Deep Learning Frameworks

Registered

2022-08-12