Torch-TensorRT v1.1.0 #1043
narendasan
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Support for PyTorch 1.11, Various Bug Fixes, New Debugging Tools, Removing Max Batch Size
Torch-TensorRT 1.1.0 targets PyTorch 1.11, CUDA 11.3, cuDNN 8.2 and TensorRT 8.2. Due to recent JetPack upgrades, this release does not support Jetson (Jetpack 5.0DP or otherwise). Jetpack 5.0DP support will arrive in a mid-cycle release (Torch-TensorRT 1.1.x) along with support for TensorRT 8.4. 1.1.0 also drops support for Python 3.6 as it has reached end of life. Following 1.0.0, this release is focused on stabilizing and improving the core of Torch-TensorRT. Many improvements have been made to the partitioning system addressing limitation many users hit while trying to partially compile PyTorch modules.
New Debugging Tools
With this release we are introducing new syntax sugar that can be used to more easily debug Torch-TensorRT compilation and execution through the use of context managers. For example, in Torch-TensorRT 1.0.0 this may be a common pattern to turn on then turn off debug info:
With Torch-TensorRT 1.1.0, this now can be done with the following code:
You can also use this API to debug the Torch-TensorRT runtime as well:
The following levels are available:
Removing Max Batch Size, Strict Types
In this release we are removing the
max_batch_size
andstrict_types
settings. These settings directly corresponded to the TensorRT settings, however were not always respected which often lead to confusion. Therefore we thought it best to disable these features as deterministic behavior could not be ensured.Porting forward from
max_batch_size
,strict_types
:max_batch_size
: The first dim in shapes provided to Torch-TensorRT are considered batch dimensions, therefore instead of settingmax_batch_size
, you can just use the Input objects directlystrict_types
: A replacement with more deterministic behavior will come with an upcoming TensorRT release.Dependencies
1.1.0 (2022-05-10)
Bug Fixes
Code Refactoring
Features
BREAKING CHANGES
Signed-off-by: Dheeraj Peri [email protected]
using torchtrtc to fail. We are renaming enabled-precisions to
enable-precision since it makes more sense as the argument can
be repeated
Signed-off-by: Naren Dasan [email protected]
Signed-off-by: Naren Dasan [email protected]
--max-batch-size
from the CLIas it has no real functional effect
Signed-off-by: Naren Dasan [email protected]
Signed-off-by: Naren Dasan [email protected]
Operators Supported
Operators Currently Supported Through Converters
Operators Currently Supported Through Evaluators
Device? device=None, bool? pin_memory=None) -> (Tensor)
Layout? layout=None, Device? device=None, bool? pin_memory=None) -> (Tensor)
Layout? layout=None, Device? device=None, bool? pin_memory=None) -> (Tensor)
This discussion was created from the release Torch-TensorRT v1.1.0.
Beta Was this translation helpful? Give feedback.
All reactions