File tree Expand file tree Collapse file tree 19 files changed +58
-58
lines changed
unittest/_torch/auto_deploy Expand file tree Collapse file tree 19 files changed +58
-58
lines changed Original file line number Diff line number Diff line change 1
1
version : " 3.9"
2
2
services :
3
3
tensorrt_llm-dev :
4
- image : urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.05 -py3-x86_64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202506271620-5539
4
+ image : urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.06 -py3-x86_64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202507101256-9530
5
5
network_mode : host
6
6
ipc : host
7
7
Original file line number Diff line number Diff line change @@ -7,7 +7,7 @@ TensorRT-LLM
7
7
[ ![ Documentation] ( https://img.shields.io/badge/docs-latest-brightgreen.svg?style=flat )] ( https://nvidia.github.io/TensorRT-LLM/ )
8
8
[ ![ python] ( https://img.shields.io/badge/python-3.12-green )] ( https://www.python.org/downloads/release/python-3123/ )
9
9
[ ![ python] ( https://img.shields.io/badge/python-3.10-green )] ( https://www.python.org/downloads/release/python-31012/ )
10
- [ ![ cuda] ( https://img.shields.io/badge/cuda-12.9.0 -green )] ( https://developer.nvidia.com/cuda-downloads )
10
+ [ ![ cuda] ( https://img.shields.io/badge/cuda-12.9.1 -green )] ( https://developer.nvidia.com/cuda-downloads )
11
11
[ ![ trt] ( https://img.shields.io/badge/TRT-10.11.0-green )] ( https://developer.nvidia.com/tensorrt )
12
12
[ ![ version] ( https://img.shields.io/badge/release-0.21.0-green )] ( ./tensorrt_llm/version.py )
13
13
[ ![ license] ( https://img.shields.io/badge/license-Apache%202-blue )] ( ./LICENSE )
Original file line number Diff line number Diff line change 1
- # These vulnerabilities were inherited from the base image (pytorch:25.05-py3) and should be removed when the base image
2
- # is updated.
3
-
4
- # WAR against https://github.com/advisories/GHSA-vqfr-h8mv-ghfj
5
- h11>=0.16.0
6
- # WAR against https://github.com/advisories/GHSA-7cx3-6m66-7c5m
7
- tornado>=6.5.0
8
- # WAR against https://github.com/advisories/GHSA-5rjg-fvgr-3xxf
9
- setuptools>=78.1.1
Original file line number Diff line number Diff line change 1
1
# Multi-stage Dockerfile
2
2
ARG BASE_IMAGE=nvcr.io/nvidia/pytorch
3
3
ARG TRITON_IMAGE=nvcr.io/nvidia/tritonserver
4
- ARG BASE_TAG=25.05 -py3
5
- ARG TRITON_BASE_TAG=25.05 -py3
4
+ ARG BASE_TAG=25.06 -py3
5
+ ARG TRITON_BASE_TAG=25.06 -py3
6
6
ARG DEVEL_IMAGE=devel
7
7
8
8
FROM ${BASE_IMAGE}:${BASE_TAG} AS base
@@ -76,15 +76,6 @@ RUN pip3 install opencv-python-headless --force-reinstall --no-deps --no-cache-d
76
76
COPY docker/common/install_deep_ep.sh install_deep_ep.sh
77
77
RUN bash ./install_deep_ep.sh && rm install_deep_ep.sh
78
78
79
- # WARs against security issues inherited from pytorch:25.04
80
- # * https://github.com/advisories/GHSA-vqfr-h8mv-ghfj
81
- # * https://github.com/advisories/GHSA-7cx3-6m66-7c5m
82
- # * https://github.com/advisories/GHSA-5rjg-fvgr-3xxf
83
- RUN pip3 install --upgrade --no-cache-dir \
84
- "h11>=0.16" \
85
- "tornado>=6.5.0" \
86
- "setuptools>=78.1.1,<80"
87
-
88
79
FROM ${TRITON_IMAGE}:${TRITON_BASE_TAG} AS triton
89
80
90
81
FROM devel AS tritondevel
Original file line number Diff line number Diff line change @@ -183,16 +183,16 @@ jenkins-aarch64_%: STAGE = tritondevel
183
183
jenkins-rockylinux8_% : IMAGE_WITH_TAG = $(shell grep '^[[:space:]]* LLM_ROCKYLINUX8_PY312_DOCKER_IMAGE = ' ../jenkins/L0_MergeRequest.groovy | grep -o '".* "' | tr -d '"')
184
184
jenkins-rockylinux8_% : STAGE = tritondevel
185
185
jenkins-rockylinux8_% : BASE_IMAGE = nvidia/cuda
186
- jenkins-rockylinux8_% : BASE_TAG = 12.9.0 -devel-rockylinux8
186
+ jenkins-rockylinux8_% : BASE_TAG = 12.9.1 -devel-rockylinux8
187
187
188
188
rockylinux8_% : STAGE = tritondevel
189
189
rockylinux8_% : BASE_IMAGE = nvidia/cuda
190
- rockylinux8_% : BASE_TAG = 12.9.0 -devel-rockylinux8
190
+ rockylinux8_% : BASE_TAG = 12.9.1 -devel-rockylinux8
191
191
192
192
# For x86_64 and aarch64
193
193
ubuntu22_% : STAGE = tritondevel
194
194
ubuntu22_% : BASE_IMAGE = nvidia/cuda
195
- ubuntu22_% : BASE_TAG = 12.9.0 -devel-ubuntu22.04
195
+ ubuntu22_% : BASE_TAG = 12.9.1 -devel-ubuntu22.04
196
196
197
197
trtllm_% : STAGE = release
198
198
trtllm_% : PUSH_TO_STAGING := 0
Original file line number Diff line number Diff line change 5
5
# This script is used for reinstalling CUDA on Rocky Linux 8 with the run file.
6
6
# CUDA version is usually aligned with the latest NGC CUDA image tag.
7
7
# Only use when public CUDA image is not ready.
8
- CUDA_VER=" 12.9.0_575.51.03 "
8
+ CUDA_VER=" 12.9.1_575.57.08 "
9
9
CUDA_VER_SHORT=" ${CUDA_VER% _* } "
10
10
11
11
NVCC_VERSION_OUTPUT=$( nvcc --version)
Original file line number Diff line number Diff line change 4
4
5
5
# Use latest stable version from https://pypi.org/project/torch/#history
6
6
# and closest to the version specified in
7
- # https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-05 .html#rel-25-05
7
+ # https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-06 .html#rel-25-06
8
8
TORCH_VERSION=" 2.7.1"
9
9
SYSTEM_ID=$( grep -oP ' (?<=^ID=).+' /etc/os-release | tr -d ' "' )
10
10
Original file line number Diff line number Diff line change @@ -5,20 +5,19 @@ set -ex
5
5
TRT_VER=" 10.11.0.33"
6
6
# Align with the pre-installed cuDNN / cuBLAS / NCCL versions from
7
7
# https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/rel-25-05.html#rel-25-05
8
- CUDA_VER=" 12.9" # 12.9.0
8
+ CUDA_VER=" 12.9" # 12.9.1
9
9
# Keep the installation for cuDNN if users want to install PyTorch with source codes.
10
10
# PyTorch 2.x can compile with cuDNN v9.
11
- CUDNN_VER=" 9.10.1.4 -1"
12
- # NCCL version 2.26.x used in the NGC PyTorch 25.05 image but has a performance regression issue.
13
- # Use NCCL version 2.27.5 which has the fixes .
11
+ CUDNN_VER=" 9.10.2.21 -1"
12
+ # PyTorch 25.06 uses NCCL 2.27.3. NCCL 2.27.5 resolves a perf regression issue.
13
+ # Use NCCL version 2.27.5 instead .
14
14
NCCL_VER=" 2.27.5-1+cuda12.9"
15
- # Use cuBLAS version 12.9.0.13 instead.
16
- CUBLAS_VER=" 12.9.0.13-1"
15
+ CUBLAS_VER=" 12.9.1.4-1"
17
16
# Align with the pre-installed CUDA / NVCC / NVRTC versions from
18
17
# https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html
19
- NVRTC_VER=" 12.9.41 -1"
20
- CUDA_RUNTIME=" 12.9.37 -1"
21
- CUDA_DRIVER_VERSION=" 575.51.03 -1.el8"
18
+ NVRTC_VER=" 12.9.86 -1"
19
+ CUDA_RUNTIME=" 12.9.79 -1"
20
+ CUDA_DRIVER_VERSION=" 575.57.08 -1.el8"
22
21
23
22
for i in " $@ " ; do
24
23
case $i in
Original file line number Diff line number Diff line change @@ -143,7 +143,7 @@ The following table shows the supported software for TensorRT-LLM.
143
143
* -
144
144
- Software Compatibility
145
145
* - Container
146
- - [25.05 ](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
146
+ - [25.06 ](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html)
147
147
* - TensorRT
148
148
- [10.11](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html)
149
149
* - Precision
Original file line number Diff line number Diff line change @@ -28,10 +28,10 @@ UPLOAD_PATH = env.uploadPath ? env.uploadPath : "sw-tensorrt-generic/llm-artifac
28
28
// Container configuration
29
29
// available tags can be found in: https://urm.nvidia.com/artifactory/sw-tensorrt-docker/tensorrt-llm/
30
30
// [base_image_name]-[arch]-[os](-[python_version])-[trt_version]-[torch_install_type]-[stage]-[date]-[mr_id]
31
- LLM_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.05 -py3-x86_64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202506271620-5539 "
32
- LLM_SBSA_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.05 -py3-aarch64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202506271620-5539 "
33
- LLM_ROCKYLINUX8_PY310_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:cuda-12.9.0 -devel-rocky8-x86_64-rocky8-py310-trt10.11.0.33-skip-tritondevel-202506271620-5539 "
34
- LLM_ROCKYLINUX8_PY312_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:cuda-12.9.0 -devel-rocky8-x86_64-rocky8-py312-trt10.11.0.33-skip-tritondevel-202506271620-5539 "
31
+ LLM_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.06 -py3-x86_64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202507101256-9530 "
32
+ LLM_SBSA_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:pytorch-25.06 -py3-aarch64-ubuntu24.04-trt10.11.0.33-skip-tritondevel-202507101256-9530 "
33
+ LLM_ROCKYLINUX8_PY310_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:cuda-12.9.1 -devel-rocky8-x86_64-rocky8-py310-trt10.11.0.33-skip-tritondevel-202507101256-9530 "
34
+ LLM_ROCKYLINUX8_PY312_DOCKER_IMAGE = " urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:cuda-12.9.1 -devel-rocky8-x86_64-rocky8-py312-trt10.11.0.33-skip-tritondevel-202507101256-9530 "
35
35
36
36
// TODO: Move common variables to an unified location
37
37
BUILD_CORES_REQUEST = " 8"
You can’t perform that action at this time.
0 commit comments