-
Sample changes
- Added 2 safety samples sampleSafeMNIST, and sampleSafePluginV3 to demonstrate how to use TensorRT with the safety workflow.
- Added trtSafeExec to accompany the safety workflow release.
- Added python/stream_writer to showcase how to serialize a TensorRT engine directly to a custom stream using the IStreamWriter interface, rather than writing to a file or a contiguous memory buffer.
- Added python/strongly_type_autocast to demonstrate how to convert FP32 ONNX models to mixed precision (FP32-FP16) using ModelOpt's AutoCast tool and subsequently building the engine with TensorRT's Strong Typing mode.
- Added sampleCudla to demonstrate how to use the cuDLA API to run TensorRT engines on the Deep Learning Accelerator (DLA) hardware, which is available on NVIDIA Jetson and DRIVE platforms.
- Deprecated sampleCharRNN.
-
Plugin changes
- Deprecated bertQKVToContextPlugin and will be removed in a future release. No alternatives are planned to be provided.
-
Parser changes
- Added support for
RotaryEmbedding,RMSNormalizationandTensorScatterfor improved LLM model support - Added more specialized quantization ops for models quantized through TensorRT ModelOptimizer.
- Added
kREPORT_CAPABILITY_DLAflag to enable per-node validation when building DLA engines through TensorRT. - Added
kENABLE_PLUGIN_OVERRIDEflag to enable TensorRT plugin override for nodes that share names with user plugins. - Improved error reporting for models with multiple subgraphs, such as
LooporScannodes.
- Added support for
-
Demo changes
- demoDiffusion: - Stable Diffusion 1.5, 2.0 and 2.1 pipelines have been deprecated and removed. - Added support for Wan2.2-T2V-A14B Text to Video pipeline
-
Sample changes
- Replace all pycuda usages with cuda-python APIs
- Removed the efficientnet samples
- Deprecated tensorflow_object_detection and efficientdet samples
- Samples will no longer be released with the packages. The TensorRT GitHub repository will be the single source.
-
Parsers:
- Added support for the
Attentionoperator - Improved refit for
ConstantOfShapenodes
- Added support for the
-
Demos
- demoDiffusion:
- Added support for the Cosmos-Predict2 text2image and video2world pipelines
- demoDiffusion:
- Added support for TensorRT API Capture and Replay feature, see the developer guide for more information.
- Demo changes
- Added support for Flux Kontext pipeline.
- Added support for CUDA 13.0, dropped support for CUDA 11.X
- Dropped support for Ubuntu 20.04
- Dropped support for Python versions < 3.10 for samples and demos
- Plugin changes
- Fixed a division-by-zero error in geluPlugin that occured when the bias is omitted.
- Completed transition away from using static plugin field/attribute member variables in standard plugins. There's no such need since presently, TRT does not access field information after plugin creators are destructed (deregistered from the plugin registry), nor does access such information without a creator instance.
- Sample changes
- Deprecated the
yolov3_onnxsample due to unstable url of yolo weights. - Updated the
1_run_onnx_with_tensorrtand2_construct_network_with_layer_apissamples to usecuda-pythoninstead ofPyCUDAfor latest GPU/CUDA support.
- Deprecated the
- Parser changes
- Decreased memory usage when importing models with external weights
- Added
loadModelProto,loadInitializerandparseModelProtoAPIs for IParser. These APIs are meant to be used to load user initializers when parsing ONNX models. - Added
loadModelProto,loadInitializerandrefitModelProtoAPIs for IParserRefitter. These APIs are meant to be used to load user initializers when refitting ONNX models. - Deprecated
IParser::parseWithWeightDescriptors.
- Plugin changes
- Migrated
IPluginV2-descendent version 1 ofcropAndResizeDynamic, to version 2, which implementsIPluginV3. - Note: The newer versions preserve the attributes and I/O of the corresponding older plugin version. The older plugin versions are deprecated and will be removed in a future release
- Deprecated the listed versions of the following plugins:
DecodeBbox3DPlugin(version 1)DetectionLayer_TRT(version 1)EfficientNMS_TRT(version 1)FlattenConcat_TRT(version 1)GenerateDetection_TRT(version 1)GridAnchor_TRT(version 1)GroupNormalizationPlugin(version 1)InstanceNormalization_TRT(version 2)ModulatedDeformConv2d(version 1)MultilevelCropAndResize_TRT(version 1)MultilevelProposeROI_TRT(version 1)RPROI_TRT(version 1)PillarScatterPlugin(version 1)PriorBox_TRT(version 1)ProposalLayer_TRT(version 1)ProposalDynamic(version 1)Region_TRT(version 1)Reorg_TRT(version 2)ResizeNearest_TRT(version 1)ScatterND(version 1)VoxelGeneratorPlugin(version 1)
- Migrated
- Demo changes
- Added Image-to-Image support for Stable Diffusion v3.5-large ControlNet models.
- Enabled download of pre-exported ONNX models for the Stable Diffusion v3.5-large pipeline.
- Sample changes
- Added two refactored python samples 1_run_onnx_with_tensorrt and 2_construct_network_with_layer_apis
- Parser changes
- Added support for integer-typed base tensors for
Powoperations - Added support for custom
MXFP8quantization operations - Added support for ellipses, diagonal, and broadcasting in
Einsumoperations
- Added support for integer-typed base tensors for
Key Features and Updates:
- Plugin changes
- Migrated
IPluginV2-descendent version 1 ofcropAndResizePluginDynamic, to version 2, which implementsIPluginV3. - Migrated
IPluginV2-descendent version 1 ofDisentangledAttention_TRT, to version 2, which implementsIPluginV3. - Migrated
IPluginV2-descendent version 1 ofMultiscaleDeformableAttnPlugin_TRT, to version 2, which implementsIPluginV3. - Note: The newer versions preserve the attributes and I/O of the corresponding older plugin version. The older plugin versions are deprecated and will be removed in a future release.
- Migrated
- Demo changes
- demoDiffusion
- Added support for Stable Diffusion 3.5-medium and 3.5-large pipelines in BF16 and FP16 precisions.
- Added support for Stable Diffusion 3.5-large pipeline in FP8 precision.
- demoDiffusion
- Parser changes
- Added
kENABLE_UINT8_AND_ASYMMETRIC_QUANTIZATION_DLAparser flag to enable UINT8 asymmetric quantization on engines targeting DLA. - Removed restriction that inputs to
RandomNormalLikeandRandomUniformLikemust be tensors. - Clarified limitations of scan outputs for
Loopnodes.
- Added
Key Features and Updates:
- Plugin changes
- Deprecated the enum classes PluginVersion & PluginCreatorVersion.
PluginVersion&PluginCreatorVersionare used only in relation toIPluginV2-descendent plugin interfaces, which are all deprecated. - Added the following APIs that enable users to obtain a list of all Plugin Creators hierarchically registered to a TensorRT
IPluginRegistry(C++,Python) instance.- C++ API:
IPluginRegistry::getAllCreatorsRecursive() - Python API:
IPluginRegistry.all_creators_recursive
- C++ API:
- Deprecated the enum classes PluginVersion & PluginCreatorVersion.
- Demo changes
- demoDiffusion
- Added FP16 and FP8 LoRA support for the SDXL and FLUX pipelines.
- Added FP16 ControlNet support for the SDXL pipeline.
- demoDiffusion
- Sample changes
- Added support for the python_plugin sample to compile targets to Blackwell.
- Parser changes
- Cleaned up log spam when the ONNX network contained a mixture of Plugins and LocalFunctions.
- UINT8 constants are now properly imported for
QuantizeLinear&DequantizeLinearnodes. - Plugin fallback importer now also reads its namespace from a Node's domain field.
Key Features and Updates:
- Demo changes
- demoDiffusion
- Added Canny ControlNet support for the SDXL pipeline
- demoDiffusion
- Plugin changes
- Added a readme to the GroupNormalization plugin (
GroupNormalizationPlugin) - 4314 - Fixed bug in
CustomQKVToConte mxtPluginDynamicversion 3 where SM 100 was not considered a supported platform.
- Added a readme to the GroupNormalization plugin (
- Parser changes
- Added support for Python AOT plugins
- Added support for opset 21 GroupNorm - 4336
- Fixed support for opset 18+ ScatterND
- Sample changes
- Added a new sample
dds_faster_rcnnwhich demonstrates how to handle data-dependent shaped outputs withIOutputAllocator.
- Added a new sample
- Fixed issues:
- Fixed streamReaderV2 Python API performance issue - 4327
Key Features and Updates:
- Demo changes
- demoDiffusion
- Added Image-to-Image support for Flux-1.dev and Flux.1-schnell pipelines.
- Added ControlNet support for FLUX.1-Canny-dev and FLUX.1-Depth-dev pipelines. Native FP8 quantization is also supported for these pipelines.
- Added support for ONNX model export only mode. See --onnx-export-only.
- Added FP16, BF16, FP8, and FP4 support for all Flux Pipelines.
- demoDiffusion
- Plugin changes
- Added SM 100 and SM 120 support to bertQKVToContextPlugin. This enables demo/BERT on Blackwell GPUs.
- Sample changes
- Added a new
sampleEditableTimingCacheto demonstrate how to build an engine with the desired tactics by modifying the timing cache. - Deleted the
sampleAlgorithmSelectorsample. - Fixed
sampleOnnxMNISTby updating the correct INT8 dynamic range.
- Added a new
- Parser changes
- Added support for
FLOAT4E2M1types for quantized networks. - Added support for dynamic axes and improved performance of
CumSumoperations. - Fixed the import of local functions when their input tensor names aliased one from an outside scope.
- Added support for
Powops with integer-typed exponent values.
- Added support for
- Fixed issues
Key Feature and Updates:
-
Demo Changes
- demoDiffusion
- Enabled low-vram for the Flux pipeline. Users can now run the pipelines on systems with 32GB VRAM.
- Added support for FLUX.1-schnell pipeline.
- Enabled weight streaming mode for Flux pipeline.
- demoDiffusion
-
Plugin Changes
- On Blackwell and later platforms, TensorRT will drop cuDNN support on the following categories of plugins
- User-written
IPluginV2Ext,IPluginV2DynamicExt, andIPluginV2IOExtplugins that are dependent on cuDNN handles provided by TensorRT (via theattachToContext()API). - TensorRT standard plugins that use cuDNN, specifically:
InstanceNormalization_TRT(version: 1, 2, and 3) present inplugin/instanceNormalizationPlugin/.GroupNormalizationPlugin(version: 1) present inplugin/groupNormalizationPlugin/.- Note: These normalization plugins are superseded by TensorRT’s native
INormalizationLayer(C++, Python). TensorRT support for cuDNN-dependent plugins remain unchanged on pre-Blackwell platforms.
- User-written
- On Blackwell and later platforms, TensorRT will drop cuDNN support on the following categories of plugins
-
Parser Changes
- Now prioritizes using plugins over local functions when a corresponding plugin is available in the registry.
- Added dynamic axes support for
SqueezeandUnsqueezeoperations. - Added support for parsing mixed-precision
BatchNormalizationnodes in strongly-typed mode.
-
Addressed Issues
- Fixed 4113.
Key Feature and Updates:
-
Demo Changes
- demoBERT: The use of
fcPluginin demoBERT has been removed. - demoBERT: All TensorRT plugins now used in demoBERT (
CustomEmbLayerNormDynamic,CustomSkipLayerNormDynamic, andCustomQKVToContextDynamic) now have versions that inherit from IPluginV3 interface classes. The user can opt-in to use these V3 plugins by specifying--use-v3-pluginsto the builder scripts.- Opting-in to use V3 plugins does not affect performance, I/O, or plugin attributes.
- There is a known issue in the V3 (version 4) of
CustomQKVToContextDynamicplugin from TensorRT 10.6.0, causing an internal assertion error if either the batch or sequence dimensions differ at runtime from the ones used to serialize the engine. See the “known issues” section of the TensorRT-10.6.0 release notes. - For smoother migration, the default behavior is still using the deprecated
IPluginV2DynamicExt-derived plugins, when the flag:--use-v3-pluginsisn't specified in the builder scripts. The flag--use-deprecated-pluginswas added as an explicit way to enforce the default behavior, and is mutually exclusive with--use-v3-plugins.
- demoDiffusion
- Introduced BF16 and FP8 support for the Flux.1-dev pipeline.
- Expanded FP8 support on Ada platforms.
- Enabled LoRA adapter compatibility for SDv1.5, SDv2.1, and SDXL pipelines using Diffusers version 0.30.3.
- demoBERT: The use of
-
Sample Changes
- Added the Python sample quickly_deployable_plugins, which demonstrates quickly deployable Python-based plugin definitions (QDPs) in TensorRT. QDPs are a simple and intuitive decorator-based approach to defining TensorRT plugins, requiring drastically less code.
-
Plugin Changes
- The
fcPluginhas been deprecated. Its functionality has been superseded by the IMatrixMultiplyLayer that is natively provided by TensorRT. - Migrated
IPluginV2-descendent version 1 ofCustomEmbLayerNormDynamic, to version 6, which implementsIPluginV3.- The newer versions preserve the attributes and I/O of the corresponding older plugin version.
- The older plugin versions are deprecated and will be removed in a future release.
- The
-
Parser Changes
- Updated ONNX submodule version to 1.17.0.
- Fixed issue where conditional layers were incorrectly being added.
- Updated local function metadata to contain more information.
- Added support for parsing nodes with Quickly Deployable Plugins.
- Fixed handling of optional outputs.
-
Tool Updates
- ONNX-Graphsurgeon updated to version 0.5.3
- Polygraphy updated to 0.49.14.
Key Features and Updates:
- Demo changes
- Added Flux.1-dev pipeline
- Sample changes
- None
- Plugin changes
- Migrated
IPluginV2-descendent versions ofbertQKVToContextPlugin(1, 2, 3) to newer versions (4, 5, 6 respectively) which implementIPluginV3. - Note:
- The newer versions preserve the attributes and I/O of the corresponding older plugin version
- The older plugin versions are deprecated and will be removed in a future release
- Migrated
- Quickstart guide
- None
- Parser changes
- Added support for real-valued
STFToperations - Improved error handling in
IParser
- Added support for real-valued
Known issues:
- Demos:
- TensorRT engine might not be build successfully when using
--fp8flag on H100 GPUs.
- TensorRT engine might not be build successfully when using
Key Features and Updates:
-
Demo changes
- Added Stable Cascade pipeline.
- Enabled INT8 and FP8 quantization for Stable Diffusion v1.5, v2.0 and v2.1 pipelines.
- Enabled FP8 quantization for Stable Diffusion XL pipeline.
-
Sample changes
- Add a new python sample
aliased_io_pluginwhich demonstrates how in-place updates to plugin inputs can be achieved through I/O aliasing.
- Add a new python sample
-
Plugin changes
- Migrated IPluginV2-descendent versions (a) of the following plugins to newer versions (b) which implement IPluginV3 (a->b):
- scatterElementsPlugin (1->2)
- skipLayerNormPlugin (1->5, 2->6, 3->7, 4->8)
- embLayerNormPlugin (2->4, 3->5)
- bertQKVToContextPlugin (1->4, 2->5, 3->6)
- Note
- The newer versions preserve the corresponding attributes and I/O of the corresponding older plugin version.
- The older plugin versions are deprecated and will be removed in a future release.
- Migrated IPluginV2-descendent versions (a) of the following plugins to newer versions (b) which implement IPluginV3 (a->b):
-
Quickstart guide
- Updated deploy_to_triton guide and removed legacy APIs.
- Removed legacy TF-TRT code as the project is no longer supported.
- Removed quantization_tutorial as pytorch_quantization has been deprecated. Check out https://github.com/NVIDIA/TensorRT-Model-Optimizer for the latest quantization support. Check Stable Diffusion XL (Base/Turbo) and Stable Diffusion 1.5 Quantization with Model Optimizer for integration with TensorRT.
-
Parser changes
- Added support for tensor
axesforPadoperations. - Added support for
BlackmanWindow,HammingWindow, andHannWindowoperations. - Improved error handling in
IParserRefitter. - Fixed kernel shape inference in multi-input convolutions.
- Added support for tensor
-
Updated tooling
- polygraphy-extension-trtexec v0.0.9
Key Features and Updates:
- Demo changes
- Added Stable Video Diffusion(
SVD) pipeline.
- Added Stable Video Diffusion(
- Plugin changes
- Deprecated Version 1 of ScatterElements plugin. It is superseded by Version 2, which implements the
IPluginV3interface.
- Deprecated Version 1 of ScatterElements plugin. It is superseded by Version 2, which implements the
- Quickstart guide
- Updated the SemanticSegmentation guide with latest APIs.
- Parser changes
- Added support for tensor
axesinputs forSlicenode. - Updated
ScatterElementsimporter to use Version 2 of ScatterElements plugin, which implements theIPluginV3interface.
- Added support for tensor
- Updated tooling
- Polygraphy v0.49.13
Key Features and Updates:
- Demo changes
- Added Stable Diffusion 3 demo.
- Plugin changes
- Version 3 of the InstanceNormalization plugin (
InstanceNormalization_TRT) has been added. This version is based on theIPluginV3interface and is used by the TensorRT ONNX parser when nativeInstanceNormalizationis disabled.
- Version 3 of the InstanceNormalization plugin (
- Tooling changes
- Pytorch Quantization development has transitioned to TensorRT Model Optimizer. All developers are encouraged to use TensorRT Model Optimizer to benefit from the latest advancements on quantization and compression.
- Build containers
- Updated default cuda versions to
12.5.0.
- Updated default cuda versions to
Key Features and Updates:
- Parser changes
- Added
supportsModelV2API - Added support for
DeformConvoperation - Added support for
PluginV3TensorRT Plugins - Marked all IParser and IParserRefitter APIs as
noexcept
- Added
- Plugin changes
- Added version 2 of ROIAlign_TRT plugin, which implements the IPluginV3 plugin interface. When importing an ONNX model with the RoiAlign op, this new version of the plugin will be inserted to the TRT network.
- Samples changes
- Added a new sample non_zero_plugin, which is a Python version of the C++ sample sampleNonZeroPlugin.
- Updated tooling
- Polygraphy v0.49.12
- ONNX-GraphSurgeon v0.5.3
Key Features and Updates:
- Parser changes
- Added support for building with
protobuf-lite. - Fixed issue when parsing and refitting models with nested
BatchNormalizationnodes. - Added support for empty inputs in custom plugin nodes.
- Added support for building with
- Demo changes
- The following demos have been removed: Jasper, Tacotron2, HuggingFace Diffusers notebook
- Updated tooling
- Polygraphy v0.49.10
- ONNX-GraphSurgeon v0.5.2
- Build Containers
- Updated default cuda versions to
12.4.0. - Added Rocky Linux 8 and Rocky Linux 9 build containers
- Updated default cuda versions to
Key Features and Updates:
- Samples changes
- Parser changes
- Added a new class
IParserRefitterthat can be used to refit a TensorRT engine with the weights of an ONNX model. kNATIVE_INSTANCENORMis now set to ON by default.- Added support for
IPluginV3interfaces from TensorRT. - Added support for
INT4quantization. - Added support for the
reductionattribute inScatterElements. - Added support for
wrappadding mode inPad
- Added a new class
- Plugin changes
- A new plugin has been added in compliance with ONNX ScatterElements.
- The TensorRT plugin library no longer has a load-time link dependency on cuBLAS or cuDNN libraries.
- All plugins which relied on cuBLAS/cuDNN handles passed through
IPluginV2Ext::attachToContext()have moved to use cuBLAS/cuDNN resources initialized by the plugin library itself. This works by dynamically loading the required cuBLAS/cuDNN library. Additionally, plugins which independently initialized their cuBLAS/cuDNN resources have also moved to dynamically loading the required library. If the respective library is not discoverable through the library path(s), these plugins will not work. - bertQKVToContextPlugin: Version 2 of this plugin now supports head sizes less than or equal to 32.
- reorgPlugin: Added a version 2 which implements IPluginV2DynamicExt.
- disentangledAttentionPlugin: Fixed a kernel bug.
- Demo changes
- HuggingFace demos have been removed. For all users using TensorRT to accelerate Large Language Model inference, please use TensorRT-LLM.
- Updated tooling
- Polygraphy v0.49.9
- ONNX-GraphSurgeon v0.5.1
- TensorRT Engine Explorer v0.1.8
- Build Containers
- RedHat/CentOS 7.x are no longer officially supported starting with TensorRT 10.0. The corresponding container has been removed from TensorRT-OSS.
Key Features and Updates:
- Demo changes
- Faster Text-to-image using SDXL & INT8 quantization using AMMO
- Updated tooling
- Polygraphy v0.49.7
Key Features and Updates:
trtexecenhancement: Added--weightlessflag to mark the engine as weightless.- Parser changes
- Added support for Hardmax operator.
- Changes to a few operator importers to ensure that TensorRT preserves the precision of operations when using strongly typed mode.
- Plugin changes
- Explicit INT8 support added to
bertQKVToContextPlugin. - Various bug fixes.
- Explicit INT8 support added to
- Updated HuggingFace demo to use transformers v4.31.0 and PyTorch v2.1.0.
Key Features and Updates:
- Update the trt_python_plugin sample.
- Python plugins API reference is part of the offical TRT Python API.
- Added samples demonstrating the usage of the progress monitor API.
- Check sampleProgressMonitor for the C++ sample.
- Check simple_progress_monitor for the Python sample.
- Remove dependencies related to python<3.8 in python samples as we no longer support python<3.8 for python samples.
- Demo changes
- Added LAMBADA dataset accuracy checks in the HuggingFace demo.
- Enabled structured sparsity and FP8 quantized batch matrix multiplication(BMM)s in attention in the NeMo demo.
- Replaced deprecated APIs in the BERT demo.
- Updated tooling
- Polygraphy v0.49.1
Key Features and Updates:
- TensorRT plugin autorhing in Python is now supported
- See the trt_python_plugin sample for reference.
- Updated default CUDA version to 12.2
- Support for BLIP models, Seq2Seq and Vision2Seq abstractions in HuggingFace demo.
- demoDiffusion refactoring and SDXL enhancements
- Additional validation asserts for NV Plugins
- Updated tooling
- TensorRT Engine Explorer v0.1.7: graph rendering for TensorRT 9.0
kgenkernels - ONNX-GraphSurgeon v0.3.29
- PyTorch quantization toolkit v2.2.0
- TensorRT Engine Explorer v0.1.7: graph rendering for TensorRT 9.0
Key Features and Updates:
-
Added the NeMo demo to demonstrate the performance benefit of using E4M3 FP8 data type with the GPT models trained with the NVIDIA NeMo Toolkit and TransformerEngine.
-
Demo Diffusion updates
- Added SDXL 1.0 txt2img pipeline
- Added ControlNet pipeline
- Huggingface demo updates
- Added Flan-T5, OPT, BLOOM, BLOOMZ, GPT-Neo, GPT-NeoX, Cerebras-GPT support with accuracy check
- Refactored code and extracted common utils into Seq2Seq class
- Optimized shape-changing overhead and achieved a >30% e2e performance gain
- Added stable KV-cache, beam search and fp16 support for all models
- Added dynamic batch size TRT inference
- Added uneven-length multi-batch inference with attention_mask support
- Added
chatcommand – interactive CLI - Upgraded PyTorch and HuggingFace version to support Hopper GPU
- Updated notebooks with much simplified demo API.
-
Added two new TensorRT samples: sampleProgressMonitor (C++) and simple_progress_reporter (Python) that are examples for using Progress Monitor during engine build.
-
The following plugins were deprecated:
BatchedNMS_TRTBatchedNMSDynamic_TRTBatchTilePlugin_TRTClip_TRTCoordConvACCropAndResizeEfficientNMS_ONNX_TRTCustomGeluPluginDynamicLReLU_TRTNMSDynamic_TRTNMS_TRTNormalize_TRTProposalSpecialSlice_TRTSplit
-
Ubuntu 18.04 has reached end of life and is no longer supported by TensorRT starting with 9.0, and the corresponding Dockerfile(s) have been removed.
-
Support for aarch64 builds will not be available in this release, and the corresponding Dockerfiles have been removed.
8.6.1 GA - 2023-05-02
TensorRT OSS release corresponding to TensorRT 8.6.1.6 GA release.
- Updates since TensorRT 8.6.0 EA release.
- Please refer to the TensorRT 8.6.1.6 GA release notes for more information.
Key Features and Updates:
- Added a new flag
--use-cuda-graphto demoDiffusion to improve performance. - Optimized GPT2 and T5 HuggingFace demos to use fp16 I/O tensors for fp16 networks.
8.6.0 EA - 2023-03-10
TensorRT OSS release corresponding to TensorRT 8.6.0.12 EA release.
- Updates since TensorRT 8.5.3 GA release.
- Please refer to the TensorRT 8.6.0.12 EA release notes for more information.
Key Features and Updates:
- demoDiffusion acceleration is now supported out of the box in TensorRT without requiring plugins.
- The following plugins have been removed accordingly: GroupNorm, LayerNorm, MultiHeadCrossAttention, MultiHeadFlashAttention, SeqLen2Spatial, and SplitGeLU.
- Added a new sample called onnx_custom_plugin.
8.5.3 GA - 2023-01-30
TensorRT OSS release corresponding to TensorRT 8.5.3.1 GA release.
- Updates since TensorRT 8.5.2 GA release.
- Please refer to the TensorRT 8.5.3 GA release notes for more information.
Key Features and Updates:
- Added the following HuggingFace demos: GPT-J-6B, GPT2-XL, and GPT2-Medium
- Added nvinfer1::plugin namespace
- Optimized KV Cache performance for T5
8.5.2 GA - 2022-12-12
TensorRT OSS release corresponding to TensorRT 8.5.2.2 GA release.
- Updates since TensorRT 8.5.1 GA release.
- Please refer to the TensorRT 8.5.2 GA release notes for more information.
Key Features and Updates:
- Plugin enhancements
- Added LayerNormPlugin, SplitGeLUPlugin, GroupNormPlugin, and SeqLen2SpatialPlugin to support stable diffusion demo.
- KV-cache and beam search to GPT2 and T5 demos
22.12 - 2022-12-06
- Stable Diffusion demo using TensorRT Plugins
- KV-cache and beam search to GPT2 and T5 demos
- Perplexity calculation to all HF demos
- Updated trex to v0.1.5
- Increased default workspace size in demoBERT to build BS=128 fp32 engines
- Use
avg_iter=8and timing cache to make demoBERT perf more stable
- None
8.5.1 GA - 2022-11-01
TensorRT OSS release corresponding to TensorRT 8.5.1.7 GA release.
- Updates since TensorRT 8.4.1 GA release.
- Please refer to the TensorRT 8.5.1 GA release notes for more information.
Key Features and Updates:
-
Samples enhancements
- Added sampleNamedDimensions which works with named dimensions.
- Updated
sampleINT8APIandintroductory_parser_samplesto useONNXmodels overCaffe/UFF - Removed UFF/Caffe samples including
sampleMNIST,end_to_end_tensorflow_mnist,sampleINT8,sampleMNISTAPI,sampleUffMNIST,sampleUffPluginV2Ext,engine_refit_mnist,int8_caffe_mnist,uff_custom_plugin,sampleFasterRCNN,sampleUffFasterRCNN,sampleGoogleNet,sampleSSD,sampleUffSSD,sampleUffMaskRCNNanduff_ssd.
-
Plugin enhancements
- Added GridAnchorRectPlugin to support rectangular feature maps in gridAnchorPlugin.
- Added ROIAlignPlugin to support the ONNX operator RoiAlign. The ONNX parser will automatically route ROIAlign ops through the plugin.
- Added Hopper support for the BERTQKVToContextPlugin plugin.
- Exposed the use_int8_scale_max attribute in the BERTQKVToContextPlugin plugin to allow users to disable the by-default usage of INT8 scale factors to optimize softmax MAX reduction in versions 2 and 3 of the plugin.
-
ONNX-TensorRT changes
- Added support for operator Reciprocal.
-
Build containers
- Updated default cuda versions to
11.8.0.
- Updated default cuda versions to
-
Tooling enhancements
- Updated onnx-graphsurgeon to v0.3.25.
- Updated Polygraphy to v0.43.1.
- Updated polygraphy-extension-trtexec to v0.0.8.
- Updated Tensorflow Quantization Toolkit to v0.2.0.
22.08 - 2022-08-16
Updated TensorRT version to 8.4.2 - see the TensorRT 8.4.2 release notes for more information
- Updated default protobuf version to 3.20.x
- Updated ONNX-TensorRT submodule version to
22.08tag - Updated
sampleIOFormatsandsampleAlgorithmSelectorto useONNXmodels overCaffe
- Fixed missing serialization member in custom
ClipPluginplugin used inuff_custom_pluginsample - Fixed various Python import issues
- Added new DeBERTA demo
- Added version 2 for
disentangledAttentionPluginto support DeBERTA v2
- None
22.07 - 2022-07-21
polygraphy-trtexec-plugintool for Polygraphy- Multi-profile support for demoBERT
- KV cache support for HF BART demo
- Updated ONNX-GS to
v0.3.20
- None
8.4.1 GA - 2022-06-14
TensorRT OSS release corresponding to TensorRT 8.4.1.5 GA release.
- Updates since TensorRT 8.2.1 GA release.
- Please refer to the TensorRT 8.4.1 GA release notes for more information.
Key Features and Updates:
-
Samples enhancements
- Added Detectron2 Mask R-CNN R50-FPN python sample
- Added a quickstart guide for NVidia Triton deployment workflow.
- Added onnx export script for sampleOnnxMnistCoordConvAC
- Removed
sampleNMT. - Removed usage of deprecated TensorRT APIs in samples.
-
EfficientDet sample
- Added support for EfficientDet Lite and AdvProp models.
- Added dynamic batch support.
- Added mixed precision engine builder.
-
HuggingFace transformer demo
- Added BART model.
- Performance speedup of GPT-2 greedy search using GPU implementation.
- Fixed GPT2 onnx export failure due to 2G file size limitation.
- Extended Megatron LayerNorm plugins to support larger hidden sizes.
- Added performance benchmarking mode.
- Enable tf32 format by default.
-
demoBERTenhancements- Add
--durationflag to perf benchmarking script. - Fixed import of
nvinfer_pluginslibrary in demoBERT on Windows.
- Add
-
Torch-QAT toolkit
quant_bert.pymodule removed. It is now upstreamed to HuggingFace QDQBERT.- Use axis0 as default for deconv.
- #1939 - Fixed path in
classification_flowexample.
-
Plugin enhancements
- Added Disentangled attention plugin,
DisentangledAttention_TRT, to support DeBERTa model. - Added Multiscale deformable attention plugin,
MultiscaleDeformableAttnPlugin_TRT, to support DDETR model. - Added new plugins: decodeBbox3DPlugin, pillarScatterPlugin, and voxelGeneratorPlugin.
- Refactored EfficientNMS plugin to support TF-TRT and implicit batch mode.
fp16support forpillarScatterPlugin.
- Added Disentangled attention plugin,
-
Build containers
- Updated default cuda versions to
11.6.2. - CentOS Linux 8 has reached End-of-Life on Dec 31, 2021. The corresponding container has been removed from TensorRT-OSS.
- Install
devtoolset-8for updated g++ versions in CentOS7 container.
- Updated default cuda versions to
-
Tooling enhancements
- Added Tensorflow Quantization Toolkit v0.1.0 for Quantization-Aware-Training of Tensorflow 2.x Keras models.
- Added TensorRT Engine Explorer v0.1.2 for inspecting TensorRT engine plans and associated inference profiling data.
- Updated Polygraphy to v0.38.0.
- Updated onnx-graphsurgeon to v0.3.19.
-
trtexecenhancements- Added
--layerPrecisionsand--layerOutputTypesflags for specifying layer-wise precision and output type constraints. - Added
--memPoolSizeflag to specify the size of workspace as well as the DLA memory pools via a unified interface. Correspondingly the--workspaceflag has been deprecated. - "End-To-End Host Latency" metric has been removed. Use the “Host Latency” metric instead. For more information, refer to Benchmarking Network section in the TensorRT Developer Guide.
- Use
enqueueV2()instead ofenqueue()when engine has explicit batch dimensions.
- Added
22.06 - 2022-06-08
- None
- Disentangled attention (DMHA) plugin refactored
- ONNX parser updated to 8.2GA
- None
22.05 - 2022-05-13
- Disentangled attention plugin for DeBERTa
- DMHA (multiscaleDeformableAttnPlugin) plugin for DDETR
- Performance benchmarking mode to HuggingFace demo
- Updated base TensorRT version to 8.2.5.1
- Updated onnx-graphsurgeon v0.3.19 CHANGELOG
- fp16 support for pillarScatterPlugin
- #1939 - Fixed path in quantization
classification_flow - Fixed GPT2 onnx export failure due to 2G limitation
- Use axis0 as default for deconv in pytorch-quantization toolkit
- Updated onnx export script for CoordConvAC sample
- Install devtoolset-8 for updated g++ version in CentOS7 container
- Usage of deprecated TensorRT APIs in samples removed
quant_bert.pymodule removed from pytorch-quantization
22.04 - 2022-04-13
- TensorRT Engine Explorer v0.1.0 README
- Detectron 2 Mask R-CNN R50-FPN python sample
- Model export script for sampleOnnxMnistCoordConvAC
- Updated base TensorRT version to 8.2.4.2
- Updated copyright headers with SPDX identifiers
- Updated onnx-graphsurgeon v0.3.17 CHANGELOG
PyramidROIAlignplugin refactor and bug fixes- Fixed
MultilevelCropAndResizecrashes on Windows - #1583 - sublicense ieee/half.h under Apache2
- Updated demo/BERT performance tables for rel-8.2
- #1774 Fix python hangs at IndexErrors when TF is imported after TensorRT
- Various bugfixes in demos - BERT, Tacotron2 and HuggingFace GPT/T5 notebooks
- Cleaned up sample READMEs
- sampleNMT removed from samples
22.03 - 2022-03-23
- EfficientDet sample enhancements
- Added support for EfficientDet Lite and AdvProp models.
- Added dynamic batch support.
- Added mixed precision engine builder.
- Better decoupling of HuggingFace demo tests
22.02 - 2022-02-04
- New plugins: decodeBbox3DPlugin, pillarScatterPlugin, and voxelGeneratorPlugin
- Extend Megatron LayerNorm plugins to support larger hidden sizes
- Refactored EfficientNMS plugin for TFTRT and added implicit batch mode support
- Update base TensorRT version to 8.2.3.0
- GPT-2 greedy search speedup - now runs on GPU
- Updates to TensorRT developer tools
- Updated ONNX parser to v8.2.3.0
- Minor updates and bugfixes
- Samples: TFOD, GPT-2, demo/BERT
- Plugins: proposalPlugin, geluPlugin, bertQKVToContextPlugin, batchedNMS
- Unused source file(s) in demo/BERT
8.2.1 GA - 2021-11-24
TensorRT OSS release corresponding to TensorRT 8.2.1.8 GA release.
-
Updates since TensorRT 8.2.0 EA release.
-
Please refer to the TensorRT 8.2.1 GA release notes for more information.
-
ONNX parser v8.2.1
- Removed duplicate constant layer checks that caused some performance regressions
- Fixed expand dynamic shape calculations
- Added parser-side checks for
Scatterlayer support
-
Sample updates
- Added Tensorflow Object Detection API converter samples, including Single Shot Detector, Faster R-CNN and Mask R-CNN models
- Multiple enhancements in HuggingFace transformer demos
- Added multi-batch support
- Fixed resultant performance regression in batchsize=1
- Fixed T5 large/T5-3B accuracy issues
- Added notebooks for T5 and GPT-2
- Added CPU benchmarking option
- Deprecated
kSTRICT_TYPES(strict type constraints). Equivalent behaviour now achieved by settingPREFER_PRECISION_CONSTRAINTS,DIRECT_IO, andREJECT_EMPTY_ALGORITHMS - Removed
sampleMovieLens - Renamed sampleReformatFreeIO to sampleIOFormats
- Add
idleTimeoption for samples to control qps - Specify default value for
precisionConstraints - Fixed reporting of TensorRT build version in trtexec
- Fixed
combineDescriptionstypo in trtexec/tracer.py - Fixed usages of of
kDIRECT_IO
-
Plugin updates
EfficientNMSplugin support extended to TF-TRT, and for clang builds.- Sanitize header definitions for BERT fused MHA plugin
- Separate C++ and cu files in
splitPluginto avoid PTX generation (required for CUDA enhanced compatibility support) - Enable C++14 build for plugins
-
ONNX tooling updates
- onnx-graphsurgeon upgraded to v0.3.14
- Polygraphy upgraded to v0.33.2
- pytorch-quantization toolkit upgraded to v2.1.2
-
Build and container fixes
- Add
SM86target to defaultGPU_ARCHSfor platforms with cuda-11.1+ - Remove deprecated
SM_35and addSM_60to defaultGPU_ARCHS - Skip CUB builds for cuda 11.0+ #1455
- Fixed cuda-10.2 container build failures in Ubuntu 20.04
- Add native ARM server build container
- Install devtoolset-8 for updated g++ version in CentOS7
- Added a note on supporting c++14 builds for CentOS7
- Fixed docker build for large UIDs #1373
- Updated README instructions for Jetpack builds
- Add
-
demo enhancements
- Updated Tacotron2 instructions and add CPU benchmarking
- Fixed issues in demoBERT python notebook
-
Documentation updates
- Updated Python documentation for
add_reduce,add_top_k, andISoftMaxLayer - Renamed default GitHub branch to
mainand updated hyperlinks
- Updated Python documentation for
8.2.0 EA - 2021-10-05
- Demo applications showcasing TensorRT inference of HuggingFace Transformers.
- Support is currently extended to GPT-2 and T5 models.
- Added support for the following ONNX operators:
EinsumIsNanGatherNDScatterScatterElementsScatterNDSignRound
- Added support for building TensorRT Python API on Windows.
- Notable API updates in TensorRT 8.2.0.6 EA release. See TensorRT Developer Guide for details.
- Added three new APIs,
IExecutionContext: getEnqueueEmitsProfile(),setEnqueueEmitsProfile(), andreportToProfiler()which can be used to collect layer profiling info when the inference is launched as a CUDA graph. - Eliminated the global logger; each
Runtime,BuilderorRefitternow has its own logger. - Added new operators:
IAssertionLayer,IConditionLayer,IEinsumLayer,IIfConditionalBoundaryLayer,IIfConditionalOutputLayer,IIfConditionalInputLayer, andIScatterLayer. - Added new
IGatherLayermodes:kELEMENTandkND - Added new
ISliceLayermodes:kFILL,kCLAMP, andkREFLECT - Added new
IUnaryLayeroperators:kSIGNandkROUND - Added new runtime class
IEngineInspectorthat can be used to inspect the detailed information of an engine, including the layer parameters, the chosen tactics, the precision used, etc. ProfilingVerbosityenums have been updated to show their functionality more explicitly.
- Added three new APIs,
- Updated TensorRT OSS container defaults to cuda 11.4
- CMake to target C++14 builds.
- Updated following ONNX operators:
GatherandGatherElementsimplementations to natively support negative indicesPadlayer to support ND padding, along withedgeandreflectpadding mode supportIflayer with general performance improvements.
- Removed
sampleMLP. - Several flags of trtexec have been deprecated:
--explicitBatchflag has been deprecated and has no effect. When the input model is in UFF or in Caffe prototxt format, the implicit batch dimension mode is used automatically; when the input model is in ONNX format, the explicit batch mode is used automatically.--explicitPrecisionflag has been deprecated and has no effect. When the input ONNX model contains Quantization/Dequantization nodes, TensorRT automatically uses explicit precision mode.--nvtxMode=[verbose|default|none]has been deprecated in favor of--profilingVerbosity=[detailed|layer_names_only|none]to show its functionality more explicitly.
21.10 - 2021-10-05
- Benchmark script for demoBERT-Megatron
- Dynamic Input Shape support for EfficientNMS plugin
- Support empty dimensions in ONNX
- INT32 and dynamic clips through elementwise in ONNX parser
- Bump TensorRT version to 8.0.3.4
- Use static shape for only single batch single sequence input in demo/BERT
- Revert to using native FC layer in demo/BERT and FCPlugin only on older GPUs.
- Update demo/Tacotron2 for TensorRT 8.0
- Updates to TensorRT developer tools
- Polygraphy v0.33.0
- Added various examples, a CLI User Guide and how-to guides.
- Added experimental support for DLA.
- Added a
data to-inputtool that can combine inputs/outputs created by--save-inputs/--save-outputs. - Added a
PluginRefRunnerwhich provides CPU reference implementations for TensorRT plugins - Made several performance improvements in the Polygraphy CUDA wrapper.
- Removed the
to-jsontool which was used to convert Pickled data generated by Polygraphy 0.26.1 and older to JSON.
- Bugfixes and documentation updates in pytorch-quantization toolkit.
- Polygraphy v0.33.0
- Bumped up package versions: tensorflow-gpu 2.5.1, pillow 8.3.2
- ONNX parser enhancements and bugfixes
- Update ONNX submodule to v1.8.0
- Update convDeconvMultiInput function to properly handle deconvs
- Update RNN documentation
- Update QDQ axis assertion
- Fix bidirectional activation alpha and beta values
- Fix opset10
Resize - Fix shape tensor unsqueeze
- Mark BOOL tiles as unsupported
- Remove unnecessary shape tensor checks
- N/A
21.09 - 2021-09-22
- Add
ONNX2TRT_VERSIONoverwrite in CMake.
- Updates to TensorRT developer tools
- Fix assertion in EfficientNMSPlugin
- N/A
21.08 - 2021-08-05
- Add demoBERT and demoBERT-MT (sparsity) benchmark data for TensorRT 8.
- Added example python notebooks
- Updated samples and plugins directory structure
- Updates to TensorRT developer tools
- README fix to update build command for native aarch64 builds.
- N/A
21.07 - 2021-07-21
Identical to the TensorRT-OSS 8.0.1 Release.
8.0.1 - 2021-07-02
- Added support for the following ONNX operators:
Celu,CumSum,EyeLike,GatherElements,GlobalLpPool,GreaterOrEqual,LessOrEqual,LpNormalization,LpPool,ReverseSequence, andSoftmaxCrossEntropyLossdetails. - Rehauled
ResizeONNX operator, now fully supporting the following modes:- Coordinate Transformation modes:
half_pixel,pytorch_half_pixel,tf_half_pixel_for_nn,asymmetric, andalign_corners. - Modes:
nearest,linear. - Nearest Modes:
floor,ceil,round_prefer_floor,round_prefer_ceil.
- Coordinate Transformation modes:
- Added support for multi-input ONNX
ConvTranposeoperator. - Added support for 3D spatial dimensions in ONNX
InstanceNormalization. - Added support for generic 2D padding in ONNX.
- ONNX
QuantizeLinearandDequantizeLinearoperators leverageIQuantizeLayerandIDequantizeLayer.- Added support for tensor scales.
- Added support for per-axis quantization.
- Added
EfficientNMS_TRT,EfficientNMS_ONNX_TRTplugins and experimental support for ONNXNonMaxSuppressionoperator. - Added
ScatterNDplugin. - Added TensorRT QuickStart Guide.
- Added new samples: engine_refit_onnx_bidaf builds an engine from ONNX BiDAF model and refits engine with new weights, efficientdet and efficientnet samples for demonstrating Object Detection using TensorRT.
- Added support for Ubuntu20.04 and RedHat/CentOS 8.3.
- Added Python 3.9 support.
- Update Polygraphy to v0.30.3.
- Update ONNX-GraphSurgeon to v0.3.10.
- Update Pytorch Quantization toolkit to v2.1.0.
- Notable TensorRT API updates
- TensorRT now declares API’s with the
noexceptkeyword. All TensorRT classes that an application inherits from (such as IPluginV2) must guarantee that methods called by TensorRT do not throw uncaught exceptions, or the behavior is undefined. - Destructors for classes with
destroy()methods were previously protected. They are now public, enabling use of smart pointers for these classes. Thedestroy()methods are deprecated.
- TensorRT now declares API’s with the
- Moved
RefitMapAPI from ONNX parser to core TensorRT. - Various bugfixes for plugins, samples and ONNX parser.
- Port demoBERT to tensorflow2 and update UFF samples to leverage nvidia-tensorflow1 container.
IPluginandIPluginFactoryinterfaces were deprecated in TensorRT 6.0 and have been removed in TensorRT 8.0. We recommend that you write new plugins or refactor existing ones to target theIPluginV2DynamicExtandIPluginV2IOExtinterfaces. For more information, refer to Migrating Plugins From TensorRT 6.x Or 7.x To TensorRT 8.x.x.- For plugins based on
IPluginV2DynamicExtandIPluginV2IOExt, certain methods with legacy function signatures (derived fromIPluginV2andIPluginV2Extbase classes) which were deprecated and marked for removal in TensorRT 8.0 will no longer be available.
- For plugins based on
- Removed
samplePluginsince it showcased IPluginExt interface, which is no longer supported in TensorRT 8.0. - Removed
sampleMovieLensandsampleMovieLensMPS. - Removed Dockerfile for Ubuntu 16.04. TensorRT 8.0 debians for Ubuntu 16.04 require python 3.5 while minimum required python version for TensorRT OSS is 3.6.
- Removed support for PowerPC builds, consistent with TensorRT GA releases.
- We had deprecated the Caffe Parser and UFF Parser in TensorRT 7.0. They are still tested and functional in TensorRT 8.0, however, we plan to remove the support in a future release. Ensure you migrate your workflow to use
tf2onnx,keras2onnxor TensorFlow-TensorRT (TF-TRT). - Refer to TensorRT 8.0.1 GA Release Notes for additional details
21.06 - 2021-06-23
- Add switch for batch-agnostic mode in NMS plugin
- Add missing model.py in
uff_custom_pluginsample
- Update to Polygraphy v0.29.2
- Update to ONNX-GraphSurgeon v0.3.9
- Fix numerical errors for float type in NMS/batchedNMS plugins
- Update demoBERT input dimensions to match Triton requirement #1051
- Optimize TLT MaskRCNN plugins:
- enable fp16 precision in multilevelCropAndResizePlugin and multilevelProposeROIPlugin
- Algorithms optimization for NMS kernels and ROIAlign kernel
- Fix invalid cuda config issue when bs is larger than 32
- Fix issues found on Jetson NANO
- Removed fcplugin from demoBERT to improve latency
21.05 - 2021-05-20
- Extended support for ONNX operator
InstanceNormalizationto 5D tensors - Support negative indices in ONNX
Gatheroperator - Add support for importing ONNX double-typed weights as float
- ONNX-GraphSurgeon (v0.3.7) support for models with externally stored weights
- Update ONNX-TensorRT to 21.05
- Relicense ONNX-TensorRT under Apache2
- demoBERT builder fixes for multi-batch
- Speedup demoBERT build using global timing cache and disable cuDNN tactics
- Standardize python package versions across OSS samples
- Bugfixes in multilevelProposeROI and bertQKV plugin
- Fix memleaks in samples logger
21.04 - 2021-04-12
- SM86 kernels for BERT MHA plugin
- Added opset13 support for
SoftMax,LogSoftmax,Squeeze, andUnsqueeze. - Added support for the
EyeLikeandGatherElementsoperators.
- Updated TensorRT version to v7.2.3.4.
- Update to ONNX-TensorRT 21.03
- ONNX-GraphSurgeon (v0.3.4) - updates fold_constants to correctly exit early.
- Set default CUDA_INSTALL_DIR #798
- Plugin bugfixes, qkv kernels for sm86
- Fixed GroupNorm CMakeFile for cu sources #1083
- Permit groupadd with non-unique GID in build containers #1091
- Avoid
reinterpret_cast#146 - Clang-format plugins and samples
- Avoid arithmetic on void pointer in multilevelProposeROIPlugin.cpp #1028
- Update BERT plugin documentation.
- Removes extra terminate call in InstanceNorm
21.03 - 2021-03-09
- Optimized FP16 NMS/batchedNMS plugins with n-bit radix sort and based on
IPluginV2DynamicExt ProposalDynamicandCropAndResizeDynamicplugins based onIPluginV2DynamicExt
- ONNX-TensorRT v21.03 update
- ONNX-GraphSurgeon v0.3.3 update
- Bugfix for
scaledSoftmaxkernel
- N/A
21.02 - 2021-02-01
- TensorRT Python API bindings
- TensorRT Python samples
- FP16 support to batchedNMSPlugin #1002
- Configurable input size for TLT MaskRCNN Plugin #986
- TensorRT version updated to 7.2.2.3
- ONNX-TensorRT v21.02 update
- Polygraphy v0.21.1 update
- PyTorch-Quantization Toolkit v2.1.0 update
- Documentation update, ONNX opset 13 support, ResNet example
- ONNX-GraphSurgeon v0.28 update
- demoBERT builder updated to work with Tensorflow2 (in compatibility mode)
- Refactor Dockerfiles for OSS container
- N/A
20.12 - 2020-12-18
- Add configurable input size for TLT MaskRCNN Plugin
- Update symbol export map for plugins
- Correctly use channel dimension when creating Prelu node
- Fix Jetson cross compilation CMakefile
- N/A
20.11 - 2020-11-20
- API documentation for ONNX-GraphSurgeon
- N/A
20.10 - 2020-10-22
- Polygraphy v0.20.13 - Deep Learning Inference Prototyping and Debugging Toolkit
- PyTorch-Quantization Toolkit v2.0.0
- Updated BERT plugins for variable sequence length inputs
- Optimized kernels for sequence lengths of 64 and 96 added
- Added Tacotron2 + Waveglow TTS demo #677
- Re-enable
GridAnchorRect_TRTplugin with rectangular feature maps #679 - Update batchedNMS plugin to IPluginV2DynamicExt interface #738
- Support 3D inputs in InstanceNormalization plugin #745
- Added this CHANGELOG.md
- ONNX GraphSurgeon - v0.2.7 with bugfixes, new examples.
- demo/BERT bugfixes for Jetson Xavier
- Updated build Dockerfile to cuda-11.1
- Updated ClangFormat style specification according to TensorRT coding guidelines
- N/A
7.2.1 - 2020-10-20
- Polygraphy v0.20.13 - Deep Learning Inference Prototyping and Debugging Toolkit
- PyTorch-Quantization Toolkit v2.0.0
- Updated BERT plugins for variable sequence length inputs
- Optimized kernels for sequence lengths of 64 and 96 added
- Added Tacotron2 + Waveglow TTS demo #677
- Re-enable
GridAnchorRect_TRTplugin with rectangular feature maps #679 - Update batchedNMS plugin to IPluginV2DynamicExt interface #738
- Support 3D inputs in InstanceNormalization plugin #745
- Added this CHANGELOG.md
- ONNX GraphSurgeon - v0.2.7 with bugfixes, new examples.
- demo/BERT bugfixes for Jetson Xavier
- Updated build Dockerfile to cuda-11.1
- Updated ClangFormat style specification according to TensorRT coding guidelines
- N/A