From 799101cb5a683eae061cb8b73387721a8cc426dd Mon Sep 17 00:00:00 2001 From: Kevin Chen Date: Mon, 13 Apr 2026 14:45:53 -0700 Subject: [PATCH] 10.16.1 update Signed-off-by: Kevin Chen --- CHANGELOG.md | 4 ++++ README.md | 18 +++++++++--------- VERSION | 2 +- docker/rockylinux8.Dockerfile | 18 +++++++++--------- docker/rockylinux9.Dockerfile | 18 +++++++++--------- docker/ubuntu-22.04.Dockerfile | 18 +++++++++--------- docker/ubuntu-24.04-aarch64.Dockerfile | 18 +++++++++--------- docker/ubuntu-24.04.Dockerfile | 18 +++++++++--------- docker/ubuntu-cross-aarch64.Dockerfile | 8 ++++---- include/NvInferVersion.h | 4 ++-- 10 files changed, 65 insertions(+), 61 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 79920ef19..08c73d95c 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,4 +1,8 @@ # TensorRT OSS Release Changelog +## 10.16.1 GA - 2026-4-13 + +- This is a bugfix release with no major new features. See the [release notes](https://docs.nvidia.com/deeplearning/tensorrt/latest/getting-started/release-notes-10/10.16.1.html) for more details. + ## 10.16 GA - 2026-3-24 - General diff --git a/README.md b/README.md index b7b7bef48..e573c9c30 100644 --- a/README.md +++ b/README.md @@ -43,7 +43,7 @@ To build the TensorRT-OSS components, you will first need the following software **TensorRT GA build** -- TensorRT v10.16.0.72 +- TensorRT v10.16.1.11 - Available from direct download links listed below **System Packages** @@ -98,24 +98,24 @@ To build the TensorRT-OSS components, you will first need the following software Else download and extract the TensorRT GA build from [NVIDIA Developer Zone](https://developer.nvidia.com) with the direct links below: - - [TensorRT 10.16.0.72 for CUDA 13.2, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz) - - [TensorRT 10.16.0.72 for CUDA 12.9, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz) - - [TensorRT 10.16.0.72 for CUDA 13.2, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/zip/TensorRT-10.16.0.72.Windows.win10.cuda-13.2.zip) - - [TensorRT 10.16.0.72 for CUDA 12.9, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/zip/TensorRT-10.16.0.72.Windows.win10.cuda-12.9.zip) + - [TensorRT 10.16.1.11 for CUDA 13.2, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz) + - [TensorRT 10.16.1.11 for CUDA 12.9, Linux x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz) + - [TensorRT 10.16.1.11 for CUDA 13.2, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/zip/TensorRT-10.16.1.11.Windows.amd64.cuda-13.2.zip) + - [TensorRT 10.16.1.11 for CUDA 12.9, Windows x86_64](https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/zip/TensorRT-10.16.1.11.Windows.amd64.cuda-12.9.zip) **Example: Ubuntu 22.04 on x86-64 with cuda-13.2** ```bash cd ~/Downloads - tar -xvzf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz - export TRT_LIBPATH=`pwd`/TensorRT-10.16.0.72/lib + tar -xvzf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz + export TRT_LIBPATH=`pwd`/TensorRT-10.16.1.11/lib ``` **Example: Windows on x86-64 with cuda-12.9** ```powershell - Expand-Archive -Path TensorRT-10.16.0.72.Windows.win10.cuda-12.9.zip - $env:TRT_LIBPATH="$pwd\TensorRT-10.16.0.72\lib" + Expand-Archive -Path TensorRT-10.16.1.11.Windows.amd64.cuda-12.9.zip + $env:TRT_LIBPATH="$pwd\TensorRT-10.16.1.11\lib" ``` ## Setting Up The Build Environment diff --git a/VERSION b/VERSION index 35e3843d3..81c6e5fbe 100644 --- a/VERSION +++ b/VERSION @@ -1 +1 @@ -10.16.0.72 +10.16.1.11 diff --git a/docker/rockylinux8.Dockerfile b/docker/rockylinux8.Dockerfile index cf7900fc4..51cf1accd 100644 --- a/docker/rockylinux8.Dockerfile +++ b/docker/rockylinux8.Dockerfile @@ -20,7 +20,7 @@ ARG CUDA_VERSION=13.2.0 FROM nvidia/cuda:${CUDA_VERSION}-devel-rockylinux8 LABEL maintainer="NVIDIA CORPORATION" -ENV TRT_VERSION 10.16.0.72 +ENV TRT_VERSION 10.16.1.11 SHELL ["/bin/bash", "-c"] # Setup user account @@ -55,15 +55,15 @@ RUN dnf install -y python38 python38-devel &&\ # Install TensorRT RUN if [ "${CUDA_VERSION:0:2}" = "13" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp38-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib64 \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp38-none-linux_x86_64.whl ;\ elif [ "${CUDA_VERSION:0:2}" = "12" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp38-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib64 \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp38-none-linux_x86_64.whl ;\ else \ echo "Invalid CUDA_VERSION"; \ exit 1; \ diff --git a/docker/rockylinux9.Dockerfile b/docker/rockylinux9.Dockerfile index d1af6aac9..b3985ba57 100644 --- a/docker/rockylinux9.Dockerfile +++ b/docker/rockylinux9.Dockerfile @@ -20,7 +20,7 @@ ARG CUDA_VERSION=13.2.0 FROM nvidia/cuda:${CUDA_VERSION}-devel-rockylinux9 LABEL maintainer="NVIDIA CORPORATION" -ENV TRT_VERSION 10.16.0.72 +ENV TRT_VERSION 10.16.1.11 SHELL ["/bin/bash", "-c"] # Setup user account @@ -60,15 +60,15 @@ RUN dnf -y install \ # Install TensorRT RUN if [ "${CUDA_VERSION:0:2}" = "13" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp39-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib64 \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp39-none-linux_x86_64.whl ;\ elif [ "${CUDA_VERSION:0:2}" = "12" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp39-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib64 \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp39-none-linux_x86_64.whl ;\ else \ echo "Invalid CUDA_VERSION"; \ exit 1; \ diff --git a/docker/ubuntu-22.04.Dockerfile b/docker/ubuntu-22.04.Dockerfile index f77203796..348c9503a 100644 --- a/docker/ubuntu-22.04.Dockerfile +++ b/docker/ubuntu-22.04.Dockerfile @@ -20,7 +20,7 @@ ARG CUDA_VERSION=13.2.0 FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04 LABEL maintainer="NVIDIA CORPORATION" -ENV TRT_VERSION 10.16.0.72 +ENV TRT_VERSION 10.16.1.11 SHELL ["/bin/bash", "-c"] # Setup user account @@ -73,15 +73,15 @@ RUN apt-get install -y --no-install-recommends \ # Install TensorRT RUN if [ "${CUDA_VERSION:0:2}" = "13" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp310-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/x86_64-linux-gnu \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp310-none-linux_x86_64.whl ;\ elif [ "${CUDA_VERSION:0:2}" = "12" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib64 \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp310-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/x86_64-linux-gnu \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp310-none-linux_x86_64.whl ;\ else \ echo "Invalid CUDA_VERSION"; \ exit 1; \ diff --git a/docker/ubuntu-24.04-aarch64.Dockerfile b/docker/ubuntu-24.04-aarch64.Dockerfile index eed97018c..f9f8076a5 100644 --- a/docker/ubuntu-24.04-aarch64.Dockerfile +++ b/docker/ubuntu-24.04-aarch64.Dockerfile @@ -20,7 +20,7 @@ ARG CUDA_VERSION=13.2.0 # Multi-arch container support available in non-cudnn containers. FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu24.04 -ENV TRT_VERSION 10.16.0.72 +ENV TRT_VERSION 10.16.1.11 SHELL ["/bin/bash", "-c"] # Setup user account and edit default account @@ -83,15 +83,15 @@ ENV PATH="/opt/venv/bin:$PATH" # Install TensorRT RUN if [ "${CUDA_VERSION:0:2}" = "13" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-13.2.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-13.2.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib/aarch64-linux-gnu/ \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp312-none-linux_aarch64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-13.2.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-13.2.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/aarch64-linux-gnu/ \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp312-none-linux_aarch64.whl ;\ elif [ "${CUDA_VERSION:0:2}" = "12" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-12.9.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-12.9.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib/aarch64-linux-gnu/ \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp312-none-linux_aarch64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-12.9.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-12.9.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/aarch64-linux-gnu/ \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp312-none-linux_aarch64.whl ;\ else \ echo "Invalid CUDA_VERSION"; \ exit 1; \ diff --git a/docker/ubuntu-24.04.Dockerfile b/docker/ubuntu-24.04.Dockerfile index bb339212e..4f63a68f9 100644 --- a/docker/ubuntu-24.04.Dockerfile +++ b/docker/ubuntu-24.04.Dockerfile @@ -21,7 +21,7 @@ ARG CUDA_VERSION=13.2.0 FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu24.04 LABEL maintainer="NVIDIA CORPORATION" -ENV TRT_VERSION=10.16.0.72 +ENV TRT_VERSION=10.16.1.11 SHELL ["/bin/bash", "-c"] # Setup user account and edit default account @@ -82,15 +82,15 @@ ENV PATH="/opt/venv/bin:$PATH" # Install TensorRT RUN if [ "${CUDA_VERSION:0:2}" = "13" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-13.2.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib/x86_64-linux-gnu/ \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp312-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-13.2.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/x86_64-linux-gnu/ \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp312-none-linux_x86_64.whl ;\ elif [ "${CUDA_VERSION:0:2}" = "12" ]; then \ - wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && tar -xf TensorRT-10.16.0.72.Linux.x86_64-gnu.cuda-12.9.tar.gz \ - && cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib/x86_64-linux-gnu/ \ - && pip install TensorRT-10.16.0.72/python/tensorrt-10.16.0.72-cp312-none-linux_x86_64.whl ;\ + wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && tar -xf TensorRT-10.16.1.11.Linux.x86_64-gnu.cuda-12.9.tar.gz \ + && cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/x86_64-linux-gnu/ \ + && pip install TensorRT-10.16.1.11/python/tensorrt-10.16.1.11-cp312-none-linux_x86_64.whl ;\ else \ echo "Invalid CUDA_VERSION"; \ exit 1; \ diff --git a/docker/ubuntu-cross-aarch64.Dockerfile b/docker/ubuntu-cross-aarch64.Dockerfile index 0994841be..4ba1daa82 100644 --- a/docker/ubuntu-cross-aarch64.Dockerfile +++ b/docker/ubuntu-cross-aarch64.Dockerfile @@ -21,7 +21,7 @@ ARG OS_VERSION=24.04 FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu${OS_VERSION} LABEL maintainer="NVIDIA CORPORATION" -ENV TRT_VERSION 10.16.0.72 +ENV TRT_VERSION 10.16.1.11 ENV DEBIAN_FRONTEND=noninteractive # Setup user account and edit default account @@ -87,9 +87,9 @@ RUN wget https://developer.download.nvidia.com/compute/cuda/13.2.0/local_install # Unpack libnvinfer. -RUN wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.0/tars/TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-13.1.tar.gz && \ - tar -xf TensorRT-10.16.0.72.Linux.aarch64-gnu.cuda-13.1.tar.gz && \ - cp -a TensorRT-10.16.0.72/lib/*.so* /usr/lib/aarch64-linux-gnu +RUN wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.16.1/tars/TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-13.2.tar.gz && \ + tar -xf TensorRT-10.16.1.11.Linux.aarch64-gnu.cuda-13.2.tar.gz && \ + cp -a TensorRT-10.16.1.11/lib/*.so* /usr/lib/aarch64-linux-gnu # Link required library RUN cd /usr/aarch64-linux-gnu/lib && ln -sf librt.so.1 librt.so diff --git a/include/NvInferVersion.h b/include/NvInferVersion.h index 212a718ca..7e2e25eaa 100644 --- a/include/NvInferVersion.h +++ b/include/NvInferVersion.h @@ -25,8 +25,8 @@ #define TRT_MAJOR_ENTERPRISE 10 #define TRT_MINOR_ENTERPRISE 16 -#define TRT_PATCH_ENTERPRISE 0 -#define TRT_BUILD_ENTERPRISE 72 +#define TRT_PATCH_ENTERPRISE 1 +#define TRT_BUILD_ENTERPRISE 11 #define NV_TENSORRT_MAJOR TRT_MAJOR_ENTERPRISE //!< TensorRT major version. #define NV_TENSORRT_MINOR TRT_MINOR_ENTERPRISE //!< TensorRT minor version. #define NV_TENSORRT_PATCH TRT_PATCH_ENTERPRISE //!< TensorRT patch version.