Skip to content

Commit 3f7abf8

Browse files
authored
[TRTLLM-6224][infra] Upgrade dependencies to DLFW 25.06 and CUDA 12.9.1 (#5678)
Signed-off-by: Yiqing Yan <[email protected]>
1 parent 4da5cfc commit 3f7abf8

File tree

18 files changed

+87
-50
lines changed

18 files changed

+87
-50
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ TensorRT-LLM
77
[![Documentation](https://img.shields.io/badge/docs-latest-brightgreen.svg?style=flat)](https://nvidia.github.io/TensorRT-LLM/)
88
[![python](https://img.shields.io/badge/python-3.12-green)](https://www.python.org/downloads/release/python-3123/)
99
[![python](https://img.shields.io/badge/python-3.10-green)](https://www.python.org/downloads/release/python-31012/)
10-
[![cuda](https://img.shields.io/badge/cuda-12.9.0-green)](https://developer.nvidia.com/cuda-downloads)
10+
[![cuda](https://img.shields.io/badge/cuda-12.9.1-green)](https://developer.nvidia.com/cuda-downloads)
1111
[![trt](https://img.shields.io/badge/TRT-10.11.0-green)](https://developer.nvidia.com/tensorrt)
1212
[![version](https://img.shields.io/badge/release-1.0.0rc5-green)](./tensorrt_llm/version.py)
1313
[![license](https://img.shields.io/badge/license-Apache%202-blue)](./LICENSE)

constraints.txt

Lines changed: 1 addition & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,5 @@
1-
# These vulnerabilities were inherited from the base image (pytorch:25.05-py3) and should be removed when the base image
1+
# These vulnerabilities were inherited from the base image (pytorch:25.06-py3) and should be removed when the base image
22
# is updated.
33

4-
# WAR against https://github.com/advisories/GHSA-vqfr-h8mv-ghfj
5-
h11>=0.16.0
6-
# WAR against https://github.com/advisories/GHSA-7cx3-6m66-7c5m
7-
tornado>=6.5.0
8-
# WAR against https://github.com/advisories/GHSA-5rjg-fvgr-3xxf
9-
setuptools>=78.1.1
104
# WAR against https://github.com/advisories/GHSA-8qvm-5x2c-j2w7
115
protobuf>=4.25.8
12-
# WAR against https://github.com/advisories/GHSA-33p9-3p43-82vq
13-
jupyter-core>=5.8.1

docker/Dockerfile.multi

Lines changed: 4 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
# Multi-stage Dockerfile
22
ARG BASE_IMAGE=nvcr.io/nvidia/pytorch
33
ARG TRITON_IMAGE=nvcr.io/nvidia/tritonserver
4-
ARG BASE_TAG=25.05-py3
5-
ARG TRITON_BASE_TAG=25.05-py3
4+
ARG BASE_TAG=25.06-py3
5+
ARG TRITON_BASE_TAG=25.06-py3
66
ARG DEVEL_IMAGE=devel
77

88
FROM ${BASE_IMAGE}:${BASE_TAG} AS base
@@ -74,18 +74,10 @@ ENV PYTORCH_CUDA_ALLOC_CONF="garbage_collection_threshold:0.99999"
7474
RUN pip3 uninstall -y opencv && rm -rf /usr/local/lib/python3*/dist-packages/cv2/
7575
RUN pip3 install opencv-python-headless --force-reinstall --no-deps --no-cache-dir
7676

77-
# WARs against security issues inherited from pytorch:25.04
78-
# * https://github.com/advisories/GHSA-vqfr-h8mv-ghfj
79-
# * https://github.com/advisories/GHSA-7cx3-6m66-7c5m
80-
# * https://github.com/advisories/GHSA-5rjg-fvgr-3xxf
77+
# WARs against security issues inherited from pytorch:25.06
8178
# * https://github.com/advisories/GHSA-8qvm-5x2c-j2w7
82-
# * https://github.com/advisories/GHSA-33p9-3p43-82vq
8379
RUN pip3 install --upgrade --no-cache-dir \
84-
"h11>=0.16" \
85-
"tornado>=6.5.0" \
86-
"setuptools>=78.1.1,<80" \
87-
"protobuf>=4.25.8" \
88-
"jupyter-core>=5.8.1"
80+
"protobuf>=4.25.8"
8981

9082
FROM ${TRITON_IMAGE}:${TRITON_BASE_TAG} AS triton
9183

docker/Makefile

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -186,16 +186,16 @@ jenkins-rockylinux8_%: PYTHON_VERSION_TAG_ID = $(if $(findstring 3.12,${PYTHON_V
186186
jenkins-rockylinux8_%: IMAGE_WITH_TAG = $(shell . ../jenkins/current_image_tags.properties && echo $$LLM_ROCKYLINUX8_${PYTHON_VERSION_TAG_ID}_DOCKER_IMAGE)
187187
jenkins-rockylinux8_%: STAGE = tritondevel
188188
jenkins-rockylinux8_%: BASE_IMAGE = nvidia/cuda
189-
jenkins-rockylinux8_%: BASE_TAG = 12.9.0-devel-rockylinux8
189+
jenkins-rockylinux8_%: BASE_TAG = 12.9.1-devel-rockylinux8
190190

191191
rockylinux8_%: STAGE = tritondevel
192192
rockylinux8_%: BASE_IMAGE = nvidia/cuda
193-
rockylinux8_%: BASE_TAG = 12.9.0-devel-rockylinux8
193+
rockylinux8_%: BASE_TAG = 12.9.1-devel-rockylinux8
194194

195195
# For x86_64 and aarch64
196196
ubuntu22_%: STAGE = tritondevel
197197
ubuntu22_%: BASE_IMAGE = nvidia/cuda
198-
ubuntu22_%: BASE_TAG = 12.9.0-devel-ubuntu22.04
198+
ubuntu22_%: BASE_TAG = 12.9.1-devel-ubuntu22.04
199199

200200
trtllm_%: STAGE = release
201201
trtllm_%: PUSH_TO_STAGING := 0

docker/common/install_base.sh

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,6 +44,8 @@ cleanup() {
4444

4545
init_ubuntu() {
4646
apt-get update
47+
# libibverbs-dev is installed but libmlx5.so is missing, reinstall the package
48+
apt-get --reinstall install -y libibverbs-dev
4749
apt-get install -y --no-install-recommends \
4850
ccache \
4951
gdb \
@@ -53,7 +55,6 @@ init_ubuntu() {
5355
llvm \
5456
libclang-rt-dev \
5557
libffi-dev \
56-
libibverbs-dev \
5758
libnuma1 \
5859
libnuma-dev \
5960
python3-dev \

docker/common/install_cuda_toolkit.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ set -ex
55
# This script is used for reinstalling CUDA on Rocky Linux 8 with the run file.
66
# CUDA version is usually aligned with the latest NGC CUDA image tag.
77
# Only use when public CUDA image is not ready.
8-
CUDA_VER="12.9.0_575.51.03"
8+
CUDA_VER="12.9.1_575.57.08"
99
CUDA_VER_SHORT="${CUDA_VER%_*}"
1010

1111
NVCC_VERSION_OUTPUT=$(nvcc --version)

docker/common/install_pytorch.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ set -ex
44

55
# Use latest stable version from https://pypi.org/project/torch/#history
66
# and closest to the version specified in
7-
# https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-05.html#rel-25-05
7+
# https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-06.html#rel-25-06
88
TORCH_VERSION="2.7.1"
99
SYSTEM_ID=$(grep -oP '(?<=^ID=).+' /etc/os-release | tr -d '"')
1010

docker/common/install_tensorrt.sh

Lines changed: 10 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,21 +4,22 @@ set -ex
44

55
TRT_VER="10.11.0.33"
66
# Align with the pre-installed cuDNN / cuBLAS / NCCL versions from
7-
# https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-05.html#rel-25-05
8-
CUDA_VER="12.9" # 12.9.0
7+
# https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-06.html#rel-25-06
8+
CUDA_VER="12.9" # 12.9.1
99
# Keep the installation for cuDNN if users want to install PyTorch with source codes.
1010
# PyTorch 2.x can compile with cuDNN v9.
11-
CUDNN_VER="9.10.1.4-1"
12-
# NCCL version 2.26.x used in the NGC PyTorch 25.05 image but has a performance regression issue.
13-
# Use NCCL version 2.27.5 which has the fixes.
11+
CUDNN_VER="9.10.2.21-1"
12+
# NGC PyTorch 25.06 image uses NCCL 2.27.3, while NCCL 2.27.5 resolves a perf regression issue.
13+
# Use NCCL version 2.27.5 instead.
1414
NCCL_VER="2.27.5-1+cuda12.9"
15-
# Use cuBLAS version 12.9.0.13 instead.
15+
# NGC PyTorch 25.06 image uses cuBLAS 12.9.1.4, but which leads to failures with MoE Lora (see https://nvbugs/5376270).
16+
# Continue using cuBLAS 12.9.0.13 until this issue is resolved.
1617
CUBLAS_VER="12.9.0.13-1"
1718
# Align with the pre-installed CUDA / NVCC / NVRTC versions from
1819
# https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html
19-
NVRTC_VER="12.9.41-1"
20-
CUDA_RUNTIME="12.9.37-1"
21-
CUDA_DRIVER_VERSION="575.51.03-1.el8"
20+
NVRTC_VER="12.9.86-1"
21+
CUDA_RUNTIME="12.9.79-1"
22+
CUDA_DRIVER_VERSION="575.57.08-1.el8"
2223

2324
for i in "$@"; do
2425
case $i in

docs/source/reference/support-matrix.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,7 @@ The following table shows the supported software for TensorRT-LLM.
153153
* -
154154
- Software Compatibility
155155
* - Container
156-
- [25.05](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
156+
- [25.06](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
157157
* - TensorRT
158158
- [10.11](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html)
159159
* - Precision

jenkins/Build.groovy

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -446,7 +446,9 @@ def runLLMBuild(pipeline, buildFlags, tarName, is_linux_x86_64)
446446
// Build tritonserver artifacts
447447
def llmPath = sh (script: "realpath ${LLM_ROOT}",returnStdout: true).trim()
448448
// TODO: Remove after the cmake version is upgraded to 3.31.8
449-
sh "cd ${LLM_ROOT}/triton_backend/inflight_batcher_llm && mkdir build && cd build && cmake .. -DTRTLLM_DIR=${llmPath} -DTRITON_COMMON_REPO_TAG=r25.05 -DTRITON_CORE_REPO_TAG=r25.05 -DTRITON_THIRD_PARTY_REPO_TAG=r25.05 -DTRITON_BACKEND_REPO_TAG=r25.05 -DUSE_CXX11_ABI=ON && make -j${BUILD_JOBS} install"
449+
// Get triton tag from docker/dockerfile.multi
450+
def tritonShortTag = sh(script: "${LLM_ROOT}/jenkins/scripts/get_triton_tag.sh ${LLM_ROOT}", returnStdout: true).trim()
451+
sh "cd ${LLM_ROOT}/triton_backend/inflight_batcher_llm && mkdir build && cd build && cmake .. -DTRTLLM_DIR=${llmPath} -DTRITON_COMMON_REPO_TAG=${tritonShortTag} -DTRITON_CORE_REPO_TAG=${tritonShortTag} -DTRITON_THIRD_PARTY_REPO_TAG=${tritonShortTag} -DTRITON_BACKEND_REPO_TAG=${tritonShortTag} -DUSE_CXX11_ABI=ON && make -j${BUILD_JOBS} install"
450452

451453
// Step 3: packaging wheels into tarfile
452454
sh "cp ${LLM_ROOT}/build/tensorrt_llm-*.whl TensorRT-LLM/"

0 commit comments

Comments
 (0)