You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CHANGELOG.md
+24
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,29 @@
1
1
# TensorRT OSS Release Changelog
2
2
3
+
## 10.8.0 GA - 2025-1-31
4
+
Key Features and Updates:
5
+
6
+
- Demo changes
7
+
- demoDiffusion
8
+
- Added [Image-to-Image](demo/Diffusion#generate-an-image-guided-by-an-initial-image-and-a-text-prompt-using-flux) support for Flux-1.dev and Flux.1-schnell pipelines.
9
+
- Added [ControlNet](demo/Diffusion#generate-an-image-guided-by-a-text-prompt-and-a-control-image-using-flux-controlnet) support for [FLUX.1-Canny-dev](https://huggingface.co/black-forest-labs/FLUX.1-Canny-dev) and [FLUX.1-Depth-dev](https://huggingface.co/black-forest-labs/FLUX.1-Depth-dev) pipelines. Native FP8 quantization is also supported for these pipelines.
10
+
- Added support for ONNX model export only mode. See [--onnx-export-only](demo/Diffusion#https://gitlab-master.nvidia.com/TensorRT/Public/oss/-/tree/release/10.8/demo/Diffusion?ref_type=heads#use-separate-directories-for-individual-onnx-models).
11
+
- Added FP16, BF16, FP8, and FP4 support for all Flux Pipelines.
12
+
- Plugin changes
13
+
- Added SM 100 and SM 120 support to bertQKVToContextPlugin. This enables demo/BERT on Blackwell GPUs.
14
+
- Sample changes
15
+
- Added a new `sampleEditableTimingCache` to demonstrate how to build an engine with the desired tactics by modifying the timing cache.
16
+
- Deleted the `sampleAlgorithmSelector` sample.
17
+
- Fixed `sampleOnnxMNIST` by updating the correct INT8 dynamic range.
18
+
- Parser changes
19
+
- Added support for `FLOAT4E2M1` types for quantized networks.
20
+
- Added support for dynamic axes and improved performance of `CumSum` operations.
21
+
- Fixed the import of local functions when their input tensor names aliased one from an outside scope.
22
+
- Added support for `Pow` ops with integer-typed exponent values.
23
+
- Fixed issues
24
+
- Fixed segmentation of boolean constant nodes - [4224](https://github.com/NVIDIA/TensorRT/issues/4224).
25
+
- Fixed accuracy issue when multiple optimization profiles were defined [4250](https://github.com/NVIDIA/TensorRT/issues/4250).
@@ -73,25 +73,25 @@ To build the TensorRT-OSS components, you will first need the following software
73
73
If using the TensorRT OSS build container, TensorRT libraries are preinstalled under `/usr/lib/x86_64-linux-gnu` and you may skip this step.
74
74
75
75
Else download and extract the TensorRT GA build from [NVIDIA Developer Zone](https://developer.nvidia.com) with the direct links below:
76
-
-[TensorRT 10.7.0.23 for CUDA 11.8, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-11.8.tar.gz)
77
-
-[TensorRT 10.7.0.23 for CUDA 12.6, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz)
78
-
-[TensorRT 10.7.0.23 for CUDA 11.8, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/zip/TensorRT-10.7.0.23.Windows.win10.cuda-11.8.zip)
79
-
-[TensorRT 10.7.0.23 for CUDA 12.6, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/zip/TensorRT-10.7.0.23.Windows.win10.cuda-12.6.zip)
76
+
-[TensorRT 10.8.0.43 for CUDA 11.8, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/tars/TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-11.8.tar.gz)
77
+
-[TensorRT 10.8.0.43 for CUDA 12.8, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/tars/TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-12.8.tar.gz)
78
+
-[TensorRT 10.8.0.43 for CUDA 11.8, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/zip/TensorRT-10.8.0.43.Windows.win10.cuda-11.8.zip)
79
+
-[TensorRT 10.8.0.43 for CUDA 12.8, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/zip/TensorRT-10.8.0.43.Windows.win10.cuda-12.8.zip)
80
80
81
81
82
-
**Example: Ubuntu 20.04 on x86-64 with cuda-12.6**
82
+
**Example: Ubuntu 20.04 on x86-64 with cuda-12.8**
83
83
84
84
```bash
85
85
cd~/Downloads
86
-
tar -xvzf TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz
87
-
export TRT_LIBPATH=`pwd`/TensorRT-10.7.0.23
86
+
tar -xvzf TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-12.8.tar.gz
@@ -101,27 +101,27 @@ For Linux platforms, we recommend that you generate a docker container for build
101
101
1. #### Generate the TensorRT-OSS build container.
102
102
The TensorRT-OSS build container can be generated using the supplied Dockerfiles and build scripts. The build containers are configured for building TensorRT OSS out-of-the-box.
103
103
104
-
**Example: Ubuntu 20.04 on x86-64 with cuda-12.6 (default)**
104
+
**Example: Ubuntu 20.04 on x86-64 with cuda-12.8 (default)**
0 commit comments