From 9880ddf72f7abc55af9c2504d47cf27a6d67a286 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 13:30:43 -0800 Subject: [PATCH 01/14] upgrade tensorrt and tensorrt_rtx --- .../scripts/generate-tensorrt-test-matrix.py | 8 ++++++++ .github/scripts/install-tensorrt-rtx.sh | 13 +++++++----- MODULE.bazel | 20 +++++++++---------- README.md | 4 ++-- dev_dep_versions.yml | 6 +++--- docsrc/getting_started/tensorrt_rtx.rst | 14 ++++++------- setup.py | 12 +++++------ third_party/tensorrt_rtx/archive/BUILD | 4 ++-- third_party/tensorrt_rtx/local/BUILD | 4 ++-- toolchains/ci_workspaces/MODULE.bazel.tmpl | 20 +++++++++---------- tools/perf/Flux/create_env.sh | 2 +- 11 files changed, 59 insertions(+), 48 deletions(-) mode change 100644 => 100755 .github/scripts/install-tensorrt-rtx.sh diff --git a/.github/scripts/generate-tensorrt-test-matrix.py b/.github/scripts/generate-tensorrt-test-matrix.py index fba74d0514..416de6596c 100644 --- a/.github/scripts/generate-tensorrt-test-matrix.py +++ b/.github/scripts/generate-tensorrt-test-matrix.py @@ -60,6 +60,10 @@ "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.2/zip/TensorRT-10.13.2.6.Windows.win10.cuda-12.9.zip", "strip_prefix": "TensorRT-10.13.2.6", }, + "10.14.1": { + "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/zip/TensorRT-10.14.1.48.Windows.win10.cuda-12.9.zip", + "strip_prefix": "TensorRT-10.14.1.48", + }, }, "linux": { "10.3.0": { @@ -94,6 +98,10 @@ "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.2/tars/TensorRT-10.13.2.6.Linux.x86_64-gnu.cuda-12.9.tar.gz", "strip_prefix": "TensorRT-10.13.2.6", }, + "10.14.1": { + "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/tars/TensorRT-10.14.1.48.Linux.x86_64-gnu.cuda-12.9.tar.gz", + "strip_prefix": "TensorRT-10.14.1.48", + }, }, } diff --git a/.github/scripts/install-tensorrt-rtx.sh b/.github/scripts/install-tensorrt-rtx.sh old mode 100644 new mode 100755 index 4e746d18c6..a1ae15c3a1 --- a/.github/scripts/install-tensorrt-rtx.sh +++ b/.github/scripts/install-tensorrt-rtx.sh @@ -1,7 +1,7 @@ install_tensorrt_rtx() { if [[ ${USE_TRT_RTX} == true ]]; then - TRT_RTX_VERSION=1.0.0.21 + TRT_RTX_VERSION=1.2.0.54 install_wheel_or_not=${1:-false} echo "It is the tensorrt-rtx build, install tensorrt-rtx with install_wheel_or_not:${install_wheel_or_not}" PLATFORM=$(python -c "import sys; print(sys.platform)") @@ -15,8 +15,8 @@ install_tensorrt_rtx() { # python version is like 3.11, we need to convert it to cp311 CPYTHON_TAG="cp${PYTHON_VERSION//./}" if [[ ${PLATFORM} == win32 ]]; then - curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-${TRT_RTX_VERSION}.Windows.win10.cuda-12.9.zip -o TensorRT-RTX-${TRT_RTX_VERSION}.Windows.win10.cuda-12.9.zip - unzip TensorRT-RTX-${TRT_RTX_VERSION}.Windows.win10.cuda-12.9.zip + curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-${TRT_RTX_VERSION}-win10-amd64-cuda-${CU_UPPERBOUND}-release-external.zip -o tensorrt-rtx-${TRT_RTX_VERSION}.win10-amd64-cuda-${CU_UPPERBOUND}.zip + unzip tensorrt-rtx-${TRT_RTX_VERSION}.win10-amd64-cuda-${CU_UPPERBOUND}.zip rtx_lib_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/lib export PATH=${rtx_lib_dir}:$PATH echo "PATH: $PATH" @@ -24,8 +24,8 @@ install_tensorrt_rtx() { pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-win_amd64.whl fi else - curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-${TRT_RTX_VERSION}.Linux.x86_64-gnu.cuda-12.9.tar.gz -o TensorRT-RTX-${TRT_RTX_VERSION}.Linux.x86_64-gnu.cuda-12.9.tar.gz - tar -xzf TensorRT-RTX-${TRT_RTX_VERSION}.Linux.x86_64-gnu.cuda-12.9.tar.gz + curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz -o tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz + tar -xzf tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz rtx_lib_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/lib export LD_LIBRARY_PATH=${rtx_lib_dir}:$LD_LIBRARY_PATH echo "LD_LIBRARY_PATH: $LD_LIBRARY_PATH" @@ -33,6 +33,9 @@ install_tensorrt_rtx() { pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-linux_x86_64.whl fi fi + # clean up the downloaded files + rm tensorrt-rtx*.tar.gz + rm -rf TensorRT-RTX-${TRT_RTX_VERSION} else echo "It is the standard tensorrt build, skip install tensorrt-rtx" fi diff --git a/MODULE.bazel b/MODULE.bazel index 80b55b9c19..31b03188bc 100644 --- a/MODULE.bazel +++ b/MODULE.bazel @@ -101,27 +101,27 @@ http_archive( http_archive( name = "tensorrt", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.x86_64-gnu.cuda-13.0.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/tars/TensorRT-10.14.1.48.Linux.x86_64-gnu.cuda-13.0.tar.gz", ], ) http_archive( name = "tensorrt_rtx", build_file = "@//third_party/tensorrt_rtx/archive:BUILD", - strip_prefix = "TensorRT-RTX-1.0.0.21", + strip_prefix = "TensorRT-RTX-1.2.0.54", urls = [ - "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-1.0.0.21.Linux.x86_64-gnu.cuda-12.9.tar.gz", + "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-1.2.0.54-linux-x86_64-cuda-13.0-release-external.tar.gz", ], ) http_archive( name = "tensorrt_sbsa", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.aarch64-gnu.cuda-13.0.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/tars/TensorRT-10.14.1.48.Linux.aarch64-gnu.cuda-13.0.tar.gz", ], ) @@ -137,18 +137,18 @@ http_archive( http_archive( name = "tensorrt_win", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/zip/TensorRT-10.13.3.9.Windows.win10.cuda-13.0.zip", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/zip/TensorRT-10.14.1.48.Windows.win10.cuda-13.0.zip", ], ) http_archive( name = "tensorrt_rtx_win", build_file = "@//third_party/tensorrt_rtx/archive:BUILD", - strip_prefix = "TensorRT-RTX-1.0.0.21", + strip_prefix = "TensorRT-RTX-1.2.0.54", urls = [ - "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-1.0.0.21.Windows.win10.cuda-12.9.zip", + "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-1.2.0.54-win10-amd64-cuda-13.0-release-external.zip", ], ) #################################################################################### diff --git a/README.md b/README.md index 4214d418bd..4a9f141481 100644 --- a/README.md +++ b/README.md @@ -7,7 +7,7 @@ Torch-TensorRT [![Documentation](https://img.shields.io/badge/docs-master-brightgreen)](https://nvidia.github.io/Torch-TensorRT/) [![pytorch](https://img.shields.io/badge/PyTorch-2.10-green)](https://download.pytorch.org/whl/nightly/cu130) [![cuda](https://img.shields.io/badge/CUDA-13.0-green)](https://developer.nvidia.com/cuda-downloads) -[![trt](https://img.shields.io/badge/TensorRT-10.13.0-green)](https://github.com/nvidia/tensorrt) +[![trt](https://img.shields.io/badge/TensorRT-10.14.0-green)](https://github.com/nvidia/tensorrt) [![license](https://img.shields.io/badge/license-BSD--3--Clause-blue)](./LICENSE) [![Linux x86-64 Nightly Wheels](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux-x86_64.yml/badge.svg?branch=nightly)](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux-x86_64.yml) [![Linux SBSA Nightly Wheels](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux-aarch64.yml/badge.svg?branch=nightly)](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux-aarch64.yml) @@ -123,7 +123,7 @@ These are the following dependencies used to verify the testcases. Torch-TensorR - Bazel 8.1.1 - Libtorch 2.10.0.dev (latest nightly) - CUDA 13.0 (CUDA 12.6 on Jetson) -- TensorRT 10.13.2.6 (TensorRT 10.3 on Jetson) +- TensorRT 10.14.1.48 (TensorRT 10.3 on Jetson) ## Deprecation Policy diff --git a/dev_dep_versions.yml b/dev_dep_versions.yml index 8f3df6e509..81b1ce714b 100644 --- a/dev_dep_versions.yml +++ b/dev_dep_versions.yml @@ -1,4 +1,4 @@ -__cuda_version__: "12.8" -__tensorrt_version__: "10.13.3" -__tensorrt_rtx_version__: "1.0.0" +__cuda_version__: "13.0" +__tensorrt_version__: "10.14.0" +__tensorrt_rtx_version__: "1.2.0" __tensorrt_llm_version__: "0.17.0.post1" diff --git a/docsrc/getting_started/tensorrt_rtx.rst b/docsrc/getting_started/tensorrt_rtx.rst index 167a4bd859..0c474fc89f 100644 --- a/docsrc/getting_started/tensorrt_rtx.rst +++ b/docsrc/getting_started/tensorrt_rtx.rst @@ -52,7 +52,7 @@ Install TensorRT-RTX Tarball ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ TensorRT-RTX tarball can be downloaded from https://developer.nvidia.com/tensorrt-rtx. -Currently, Torch-TensorRT uses TensorRT-RTX version **1.0.0.21**. +Currently, Torch-TensorRT uses TensorRT-RTX version **1.2.0.54**. Once downloaded: @@ -62,8 +62,8 @@ Make sure you add the lib path to the ``LD_LIBRARY_PATH`` environment variable. .. code-block:: sh - # If TensorRT-RTX is downloaded in /your_local_download_path/TensorRT-RTX-1.0.0.21 - export LD_LIBRARY_PATH=/your_local_download_path/TensorRT-RTX-1.0.0.21/lib:$LD_LIBRARY_PATH + # If TensorRT-RTX is downloaded in /your_local_download_path/TensorRT-RTX-1.2.0.54 + export LD_LIBRARY_PATH=/your_local_download_path/TensorRT-RTX-1.2.0.54/lib:$LD_LIBRARY_PATH echo $LD_LIBRARY_PATH | grep TensorRT-RTX **In Windows:** @@ -72,8 +72,8 @@ Make sure you add the lib path to the Windows system variable ``PATH``. .. code-block:: sh - # If TensorRT-RTX is downloaded in C:\your_local_download_path\TensorRT-RTX-1.0.0.21 - set PATH="%PATH%;C:\your_local_download_path\TensorRT-RTX-1.0.0.21\lib" + # If TensorRT-RTX is downloaded in C:\your_local_download_path\TensorRT-RTX-1.2.0.54 + set PATH="%PATH%;C:\your_local_download_path\TensorRT-RTX-1.2.0.54\lib" echo %PATH% | findstr TensorRT-RTX Install TensorRT-RTX Wheel @@ -84,8 +84,8 @@ You must install it manually from the downloaded tarball. .. code-block:: sh - # If the tarball is downloaded in /your_local_download_path/TensorRT-RTX-1.0.0.21 - python -m pip install /your_local_download_path/TensorRT-RTX-1.0.0.21/python/tensorrt_rtx-1.0.0.21-cp39-none-linux_x86_64.whl + # If the tarball is downloaded in /your_local_download_path/TensorRT-RTX-1.2.0.54 + python -m pip install /your_local_download_path/TensorRT-RTX-1.2.0.54/python/tensorrt_rtx-1.2.0.54-cp39-none-linux_x86_64.whl Build Torch-TensorRT with TensorRT-RTX -------------------------------------- diff --git a/setup.py b/setup.py index 2a5f65c61a..de000a6973 100644 --- a/setup.py +++ b/setup.py @@ -754,14 +754,14 @@ def get_requirements(): requirements = requirements + ["torch>=2.10.0.dev,<2.11.0"] if USE_TRT_RTX: requirements = requirements + [ - "tensorrt_rtx>=1.0.0.21", + "tensorrt_rtx>=1.2.0.54", ] else: requirements = requirements + [ - "tensorrt>=10.13.0,<10.14.0", - f"{tensorrt_prefix}>=10.13.0,<10.14.0", - f"{tensorrt_prefix}-bindings>=10.13.0,<10.14.0", - f"{tensorrt_prefix}-libs>=10.13.0,<10.14.0", + "tensorrt>=10.14.0,<10.15.0", + f"{tensorrt_prefix}>=10.14.0,<10.15.0", + f"{tensorrt_prefix}-bindings>=10.14.0,<10.15.0", + f"{tensorrt_prefix}-libs>=10.14.0,<10.15.0", ] return requirements @@ -781,7 +781,7 @@ def get_sbsa_requirements(): # also due to we use sbsa torch_tensorrt wheel for thor, so when we build sbsa wheel, we need to only include tensorrt dependency. return sbsa_requirements + [ "torch>=2.10.0.dev,<2.11.0", - "tensorrt>=10.13.0,<10.14.0", + "tensorrt>=10.14.0,<10.15.0", ] diff --git a/third_party/tensorrt_rtx/archive/BUILD b/third_party/tensorrt_rtx/archive/BUILD index ec6ebbe985..3078aaaf2f 100644 --- a/third_party/tensorrt_rtx/archive/BUILD +++ b/third_party/tensorrt_rtx/archive/BUILD @@ -38,7 +38,7 @@ cc_library( cc_import( name = "nvinfer_lib", shared_library = select({ - ":rtx_win": "lib/tensorrt_rtx_1_0.dll", + ":rtx_win": "lib/tensorrt_rtx_1_2.dll", ":rtx_x86_64": "lib/libtensorrt_rtx.so", }), visibility = ["//visibility:private"], @@ -47,7 +47,7 @@ cc_import( cc_import( name = "nvinfer_static_lib", static_library = select({ - ":rtx_win": "lib/tensorrt_rtx_1_0.lib", + ":rtx_win": "lib/tensorrt_rtx_1_2.lib", }), visibility = ["//visibility:private"], ) diff --git a/third_party/tensorrt_rtx/local/BUILD b/third_party/tensorrt_rtx/local/BUILD index fe844170fb..1ec61cf52e 100644 --- a/third_party/tensorrt_rtx/local/BUILD +++ b/third_party/tensorrt_rtx/local/BUILD @@ -50,7 +50,7 @@ cc_library( cc_import( name = "nvinfer_static_lib", static_library = select({ - ":rtx_win": "lib/tensorrt_rtx_1_0.lib", + ":rtx_win": "lib/tensorrt_rtx_1_2.lib", }), visibility = ["//visibility:private"], ) @@ -58,7 +58,7 @@ cc_import( cc_import( name = "nvinfer_lib", shared_library = select({ - ":rtx_win": "lib/tensorrt_rtx_1_0.dll", + ":rtx_win": "bin/tensorrt_rtx_1_2.dll", ":rtx_x86_64": "lib/libtensorrt_rtx.so", }), visibility = ["//visibility:private"], diff --git a/toolchains/ci_workspaces/MODULE.bazel.tmpl b/toolchains/ci_workspaces/MODULE.bazel.tmpl index e8066e97e6..f3dcf3ccd4 100644 --- a/toolchains/ci_workspaces/MODULE.bazel.tmpl +++ b/toolchains/ci_workspaces/MODULE.bazel.tmpl @@ -75,27 +75,27 @@ http_archive = use_repo_rule("@bazel_tools//tools/build_defs/repo:http.bzl", "ht http_archive( name = "tensorrt", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.x86_64-gnu.cuda-${CU_UPPERBOUND}.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/tars/TensorRT-10.14.1.48.Linux.x86_64-gnu.cuda-${CU_UPPERBOUND}.tar.gz", ], ) http_archive( name = "tensorrt_rtx", build_file = "@//third_party/tensorrt_rtx/archive:BUILD", - strip_prefix = "TensorRT-RTX-1.0.0.21", + strip_prefix = "TensorRT-RTX-1.2.0.54", urls = [ - "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-1.0.0.21.Linux.x86_64-gnu.cuda-12.9.tar.gz", + "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-1.2.0.54-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz", ], ) http_archive( name = "tensorrt_sbsa", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.aarch64-gnu.cuda-13.0.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/tars/TensorRT-10.14.1.48.Linux.aarch64-gnu.cuda-13.0.tar.gz", ], ) @@ -111,18 +111,18 @@ http_archive( http_archive( name = "tensorrt_win", build_file = "@//third_party/tensorrt/archive:BUILD", - strip_prefix = "TensorRT-10.13.3.9", + strip_prefix = "TensorRT-10.14.1.48", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/zip/TensorRT-10.13.3.9.Windows.win10.cuda-${CU_UPPERBOUND}.zip", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.14.1/zip/TensorRT-10.14.1.48.Windows.win10.cuda-${CU_UPPERBOUND}.zip", ], ) http_archive( name = "tensorrt_rtx_win", build_file = "@//third_party/tensorrt_rtx/archive:BUILD", - strip_prefix = "TensorRT-RTX-1.0.0.21", + strip_prefix = "TensorRT-RTX-1.2.0.54", urls = [ - "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.0/TensorRT-RTX-1.0.0.21.Windows.win10.cuda-12.9.zip", + "https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-1.2.0.54-win10-amd64-cuda-${CU_UPPERBOUND}-release-external.zip", ], ) diff --git a/tools/perf/Flux/create_env.sh b/tools/perf/Flux/create_env.sh index 330e6c53d6..ce9b849cdb 100644 --- a/tools/perf/Flux/create_env.sh +++ b/tools/perf/Flux/create_env.sh @@ -15,7 +15,7 @@ bazel cd /home/TensorRT python -m pip install --pre -e . --extra-index-url https://download.pytorch.org/whl/nightly/cu130 -pip install tensorrt==10.13.2.6 --force-reinstall +pip install tensorrt==10.14.1.48 --force-reinstall pip3 install --pre torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cu130 From d734b6875f0762b14d4c52650ceac5da6f3eb731 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 13:35:49 -0800 Subject: [PATCH 02/14] test --- .github/scripts/install-tensorrt-rtx.sh | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/.github/scripts/install-tensorrt-rtx.sh b/.github/scripts/install-tensorrt-rtx.sh index a1ae15c3a1..52d5dd7c05 100755 --- a/.github/scripts/install-tensorrt-rtx.sh +++ b/.github/scripts/install-tensorrt-rtx.sh @@ -1,6 +1,11 @@ install_tensorrt_rtx() { if [[ ${USE_TRT_RTX} == true ]]; then + if [[ ${CU_VERSION:2:2} == "13" ]]; then + export CU_UPPERBOUND="13.0" + else + export CU_UPPERBOUND="12.9" + fi TRT_RTX_VERSION=1.2.0.54 install_wheel_or_not=${1:-false} echo "It is the tensorrt-rtx build, install tensorrt-rtx with install_wheel_or_not:${install_wheel_or_not}" From 66f43e607fe9c7c1fa0a74a0f51579ec0961423b Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 14:31:12 -0800 Subject: [PATCH 03/14] test --- .github/scripts/install-tensorrt-rtx.sh | 3 +-- setup.py | 3 ++- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/scripts/install-tensorrt-rtx.sh b/.github/scripts/install-tensorrt-rtx.sh index 52d5dd7c05..43ba07159e 100755 --- a/.github/scripts/install-tensorrt-rtx.sh +++ b/.github/scripts/install-tensorrt-rtx.sh @@ -38,9 +38,8 @@ install_tensorrt_rtx() { pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-linux_x86_64.whl fi fi - # clean up the downloaded files + # clean up the downloaded rtx tarball rm tensorrt-rtx*.tar.gz - rm -rf TensorRT-RTX-${TRT_RTX_VERSION} else echo "It is the standard tensorrt build, skip install tensorrt-rtx" fi diff --git a/setup.py b/setup.py index de000a6973..54b2def76b 100644 --- a/setup.py +++ b/setup.py @@ -758,7 +758,8 @@ def get_requirements(): ] else: requirements = requirements + [ - "tensorrt>=10.14.0,<10.15.0", + # directly use tensorrt>=10.14.0,<10.15.0 in cu12* env, it will pull both tensorrt_cu12 and tensorrt_cu13 + # which will cause the conflict due to cuda-toolkit 13 is also pulled in, so we need to specify tensorrt_cu12 or tensorrt_cu13 here f"{tensorrt_prefix}>=10.14.0,<10.15.0", f"{tensorrt_prefix}-bindings>=10.14.0,<10.15.0", f"{tensorrt_prefix}-libs>=10.14.0,<10.15.0", From c25e91f983b333fe7df732eb4ebd52803e31d1e8 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 15:01:44 -0800 Subject: [PATCH 04/14] test --- setup.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/setup.py b/setup.py index 54b2def76b..2db2f97dd0 100644 --- a/setup.py +++ b/setup.py @@ -758,11 +758,11 @@ def get_requirements(): ] else: requirements = requirements + [ - # directly use tensorrt>=10.14.0,<10.15.0 in cu12* env, it will pull both tensorrt_cu12 and tensorrt_cu13 + # directly use tensorrt>=10.14.1,<10.15.0 in cu12* env, it will pull both tensorrt_cu12 and tensorrt_cu13 # which will cause the conflict due to cuda-toolkit 13 is also pulled in, so we need to specify tensorrt_cu12 or tensorrt_cu13 here - f"{tensorrt_prefix}>=10.14.0,<10.15.0", - f"{tensorrt_prefix}-bindings>=10.14.0,<10.15.0", - f"{tensorrt_prefix}-libs>=10.14.0,<10.15.0", + f"{tensorrt_prefix}>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", ] return requirements @@ -782,7 +782,7 @@ def get_sbsa_requirements(): # also due to we use sbsa torch_tensorrt wheel for thor, so when we build sbsa wheel, we need to only include tensorrt dependency. return sbsa_requirements + [ "torch>=2.10.0.dev,<2.11.0", - "tensorrt>=10.14.0,<10.15.0", + "tensorrt>=10.14.1,<10.15.0", ] From 6d9dfe0e316794881bc8e76fc12c4e9ec74b8097 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 20:56:45 -0800 Subject: [PATCH 05/14] test --- .github/scripts/install-tensorrt-rtx.sh | 6 ++++-- third_party/tensorrt/archive/BUILD | 2 +- third_party/tensorrt/local/BUILD | 2 +- 3 files changed, 6 insertions(+), 4 deletions(-) diff --git a/.github/scripts/install-tensorrt-rtx.sh b/.github/scripts/install-tensorrt-rtx.sh index 43ba07159e..ef5a2068ed 100755 --- a/.github/scripts/install-tensorrt-rtx.sh +++ b/.github/scripts/install-tensorrt-rtx.sh @@ -28,6 +28,8 @@ install_tensorrt_rtx() { if [[ ${install_wheel_or_not} == true ]]; then pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-win_amd64.whl fi + # clean up the downloaded rtx zip + rm tensorrt-rtx*.zip else curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz -o tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz tar -xzf tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz @@ -37,9 +39,9 @@ install_tensorrt_rtx() { if [[ ${install_wheel_or_not} == true ]]; then pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-linux_x86_64.whl fi + # clean up the downloaded rtx tarball + rm tensorrt-rtx*.tar.gz fi - # clean up the downloaded rtx tarball - rm tensorrt-rtx*.tar.gz else echo "It is the standard tensorrt build, skip install tensorrt-rtx" fi diff --git a/third_party/tensorrt/archive/BUILD b/third_party/tensorrt/archive/BUILD index 061947eb62..c921a7149c 100644 --- a/third_party/tensorrt/archive/BUILD +++ b/third_party/tensorrt/archive/BUILD @@ -38,7 +38,7 @@ cc_library( cc_import( name = "nvinfer_lib", shared_library = select({ - ":windows": "lib/nvinfer_10.dll", + ":windows": "bin/nvinfer_10.dll", "//conditions:default": "lib/libnvinfer.so", }), visibility = ["//visibility:private"], diff --git a/third_party/tensorrt/local/BUILD b/third_party/tensorrt/local/BUILD index d608063529..b28ef63e7c 100644 --- a/third_party/tensorrt/local/BUILD +++ b/third_party/tensorrt/local/BUILD @@ -95,7 +95,7 @@ cc_import( shared_library = select({ ":aarch64_linux": "lib/aarch64-linux-gnu/libnvinfer.so", ":ci_rhel_x86_64_linux": "lib64/libnvinfer.so", - ":windows": "lib/nvinfer_10.dll", + ":windows": "bin/nvinfer_10.dll", "//conditions:default": "lib/x86_64-linux-gnu/libnvinfer.so", }), visibility = ["//visibility:private"], From 80800e4eb3ee5d7064e51db60d395f1bd4dd12d2 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Thu, 13 Nov 2025 23:17:21 -0800 Subject: [PATCH 06/14] test --- setup.py | 29 +++++++++++++++-------------- 1 file changed, 15 insertions(+), 14 deletions(-) diff --git a/setup.py b/setup.py index 2db2f97dd0..3b8bbead9a 100644 --- a/setup.py +++ b/setup.py @@ -733,14 +733,6 @@ def run(self): "nvidia-cuda-runtime-cu13==0.0.0a0", ] -cuda_version = torch.version.cuda -if cuda_version.startswith("12"): - tensorrt_prefix = "tensorrt-cu12" -elif cuda_version.startswith("13"): - tensorrt_prefix = "tensorrt-cu13" -else: - raise ValueError(f"Unsupported CUDA version: {cuda_version}") - def get_requirements(): if IS_JETPACK: @@ -757,13 +749,22 @@ def get_requirements(): "tensorrt_rtx>=1.2.0.54", ] else: - requirements = requirements + [ + cuda_version = torch.version.cuda + if cuda_version.startswith("12"): # directly use tensorrt>=10.14.1,<10.15.0 in cu12* env, it will pull both tensorrt_cu12 and tensorrt_cu13 - # which will cause the conflict due to cuda-toolkit 13 is also pulled in, so we need to specify tensorrt_cu12 or tensorrt_cu13 here - f"{tensorrt_prefix}>=10.14.1,<10.15.0", - f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0", - f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", - ] + # which will cause the conflict due to cuda-toolkit 13 is also pulled in, so we need to specify tensorrt_cu12 here + tensorrt_prefix = "tensorrt-cu12" + requirements = requirements + [ + f"{tensorrt_prefix}>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", + ] + elif cuda_version.startswith("13"): + requirements = requirements + [ + "tensorrt>=10.14.1,<10.15.0", + ] + else: + raise ValueError(f"Unsupported CUDA version: {cuda_version}") return requirements From b69ba409b51d38c913cca26c1149bf59f23ce175 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 03:44:20 -0800 Subject: [PATCH 07/14] test --- setup.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/setup.py b/setup.py index 3b8bbead9a..c61bbdcf44 100644 --- a/setup.py +++ b/setup.py @@ -760,8 +760,11 @@ def get_requirements(): f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", ] elif cuda_version.startswith("13"): + tensorrt_prefix = "tensorrt-cu13" requirements = requirements + [ - "tensorrt>=10.14.1,<10.15.0", + f"{tensorrt_prefix}==10.14.1.48.post1", + f"{tensorrt_prefix}-bindings==10.14.1.48.post1", + f"{tensorrt_prefix}-libs==10.14.1.48.post1", ] else: raise ValueError(f"Unsupported CUDA version: {cuda_version}") From 5c61ee00059fb6fa66efafca0a9d999f9d32ad03 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 10:04:18 -0800 Subject: [PATCH 08/14] test --- .github/workflows/build_linux.yml | 2 +- setup.py | 6 +++--- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/.github/workflows/build_linux.yml b/.github/workflows/build_linux.yml index 007f23b7c2..16423c9c4c 100644 --- a/.github/workflows/build_linux.yml +++ b/.github/workflows/build_linux.yml @@ -351,7 +351,7 @@ jobs: if [[ ${{ inputs.architecture }} == "aarch64" ]]; then echo "Skipping smoke test for aarch64, since it is not an actual gpu runner" else - ${CONDA_RUN} pip install "${{ inputs.repository }}/dist/$WHEEL_NAME" + ${CONDA_RUN} pip install "${{ inputs.repository }}/dist/$WHEEL_NAME" --extra-index-url https://download.pytorch.org/${CHANNEL}/${CU_VERSION} # Checking that we have a pinned version of torch in our dependency tree ( pushd "${RUNNER_TEMP}" diff --git a/setup.py b/setup.py index c61bbdcf44..a576f607ba 100644 --- a/setup.py +++ b/setup.py @@ -762,9 +762,9 @@ def get_requirements(): elif cuda_version.startswith("13"): tensorrt_prefix = "tensorrt-cu13" requirements = requirements + [ - f"{tensorrt_prefix}==10.14.1.48.post1", - f"{tensorrt_prefix}-bindings==10.14.1.48.post1", - f"{tensorrt_prefix}-libs==10.14.1.48.post1", + f"{tensorrt_prefix}>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0", + f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", ] else: raise ValueError(f"Unsupported CUDA version: {cuda_version}") From e0e3fd6dcc692b4375a5aa177638e9f06ad8d78a Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 14:11:29 -0800 Subject: [PATCH 09/14] modify --- setup.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/setup.py b/setup.py index a576f607ba..d151e0df0c 100644 --- a/setup.py +++ b/setup.py @@ -762,9 +762,9 @@ def get_requirements(): elif cuda_version.startswith("13"): tensorrt_prefix = "tensorrt-cu13" requirements = requirements + [ - f"{tensorrt_prefix}>=10.14.1,<10.15.0", - f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0", - f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0", + f"{tensorrt_prefix}>=10.14.1,<10.15.0,!=10.14.1.48", + f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0,!=10.14.1.48", + f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0,!=10.14.1.48", ] else: raise ValueError(f"Unsupported CUDA version: {cuda_version}") From aa9769c7c3a1f0415a9537a58a03818c8074e812 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 15:01:53 -0800 Subject: [PATCH 10/14] modify --- setup.py | 1 + 1 file changed, 1 insertion(+) diff --git a/setup.py b/setup.py index d151e0df0c..c4c30ea571 100644 --- a/setup.py +++ b/setup.py @@ -765,6 +765,7 @@ def get_requirements(): f"{tensorrt_prefix}>=10.14.1,<10.15.0,!=10.14.1.48", f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0,!=10.14.1.48", f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0,!=10.14.1.48", + "nvidia-cuda-runtime==13.0.96", ] else: raise ValueError(f"Unsupported CUDA version: {cuda_version}") From dae648a5f48c38ec6283c3d074e13503652fa276 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 15:23:47 -0800 Subject: [PATCH 11/14] fix rtx issue --- .github/workflows/build_linux.yml | 2 +- setup.py | 1 - third_party/tensorrt_rtx/archive/BUILD | 2 +- 3 files changed, 2 insertions(+), 3 deletions(-) diff --git a/.github/workflows/build_linux.yml b/.github/workflows/build_linux.yml index 16423c9c4c..995865f6e3 100644 --- a/.github/workflows/build_linux.yml +++ b/.github/workflows/build_linux.yml @@ -351,7 +351,7 @@ jobs: if [[ ${{ inputs.architecture }} == "aarch64" ]]; then echo "Skipping smoke test for aarch64, since it is not an actual gpu runner" else - ${CONDA_RUN} pip install "${{ inputs.repository }}/dist/$WHEEL_NAME" --extra-index-url https://download.pytorch.org/${CHANNEL}/${CU_VERSION} + ${CONDA_RUN} pip install "${{ inputs.repository }}/dist/$WHEEL_NAME" --use-deprecated=legacy-resolver --extra-index-url https://download.pytorch.org/${CHANNEL}/${CU_VERSION} # Checking that we have a pinned version of torch in our dependency tree ( pushd "${RUNNER_TEMP}" diff --git a/setup.py b/setup.py index c4c30ea571..d151e0df0c 100644 --- a/setup.py +++ b/setup.py @@ -765,7 +765,6 @@ def get_requirements(): f"{tensorrt_prefix}>=10.14.1,<10.15.0,!=10.14.1.48", f"{tensorrt_prefix}-bindings>=10.14.1,<10.15.0,!=10.14.1.48", f"{tensorrt_prefix}-libs>=10.14.1,<10.15.0,!=10.14.1.48", - "nvidia-cuda-runtime==13.0.96", ] else: raise ValueError(f"Unsupported CUDA version: {cuda_version}") diff --git a/third_party/tensorrt_rtx/archive/BUILD b/third_party/tensorrt_rtx/archive/BUILD index 3078aaaf2f..9be0b2f5d8 100644 --- a/third_party/tensorrt_rtx/archive/BUILD +++ b/third_party/tensorrt_rtx/archive/BUILD @@ -38,7 +38,7 @@ cc_library( cc_import( name = "nvinfer_lib", shared_library = select({ - ":rtx_win": "lib/tensorrt_rtx_1_2.dll", + ":rtx_win": "bin/tensorrt_rtx_1_2.dll", ":rtx_x86_64": "lib/libtensorrt_rtx.so", }), visibility = ["//visibility:private"], From d9284fede15f6934cad89329adf780332893ac14 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 17:18:12 -0800 Subject: [PATCH 12/14] modify --- .github/scripts/install-tensorrt-rtx.sh | 6 ++++-- .github/scripts/install-torch-tensorrt.sh | 2 +- py/torch_tensorrt/_utils.py | 7 ++++++- tests/py/dynamo/conversion/test_pdist_aten.py | 16 +++++++++------- 4 files changed, 20 insertions(+), 11 deletions(-) diff --git a/.github/scripts/install-tensorrt-rtx.sh b/.github/scripts/install-tensorrt-rtx.sh index ef5a2068ed..dd2c038c40 100755 --- a/.github/scripts/install-tensorrt-rtx.sh +++ b/.github/scripts/install-tensorrt-rtx.sh @@ -23,7 +23,8 @@ install_tensorrt_rtx() { curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-${TRT_RTX_VERSION}-win10-amd64-cuda-${CU_UPPERBOUND}-release-external.zip -o tensorrt-rtx-${TRT_RTX_VERSION}.win10-amd64-cuda-${CU_UPPERBOUND}.zip unzip tensorrt-rtx-${TRT_RTX_VERSION}.win10-amd64-cuda-${CU_UPPERBOUND}.zip rtx_lib_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/lib - export PATH=${rtx_lib_dir}:$PATH + rtx_bin_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/bin + export PATH=${rtx_lib_dir}:${rtx_bin_dir}:$PATH echo "PATH: $PATH" if [[ ${install_wheel_or_not} == true ]]; then pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-win_amd64.whl @@ -34,7 +35,8 @@ install_tensorrt_rtx() { curl -L https://developer.nvidia.com/downloads/trt/rtx_sdk/secure/1.2/tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz -o tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz tar -xzf tensorrt-rtx-${TRT_RTX_VERSION}-linux-x86_64-cuda-${CU_UPPERBOUND}-release-external.tar.gz rtx_lib_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/lib - export LD_LIBRARY_PATH=${rtx_lib_dir}:$LD_LIBRARY_PATH + rtx_bin_dir=${PWD}/TensorRT-RTX-${TRT_RTX_VERSION}/bin + export LD_LIBRARY_PATH=${rtx_lib_dir}:${rtx_bin_dir}:$LD_LIBRARY_PATH echo "LD_LIBRARY_PATH: $LD_LIBRARY_PATH" if [[ ${install_wheel_or_not} == true ]]; then pip install TensorRT-RTX-${TRT_RTX_VERSION}/python/tensorrt_rtx-${TRT_RTX_VERSION}-${CPYTHON_TAG}-none-linux_x86_64.whl diff --git a/.github/scripts/install-torch-tensorrt.sh b/.github/scripts/install-torch-tensorrt.sh index 7d0b7a5947..dccf4925a2 100755 --- a/.github/scripts/install-torch-tensorrt.sh +++ b/.github/scripts/install-torch-tensorrt.sh @@ -32,7 +32,7 @@ fi if [[ ${PLATFORM} == win32 ]]; then pip install ${RUNNER_ARTIFACT_DIR}/torch_tensorrt*.whl else - pip install /opt/torch-tensorrt-builds/torch_tensorrt*.whl + pip install /opt/torch-tensorrt-builds/torch_tensorrt*.whl --use-deprecated=legacy-resolver fi echo -e "Running test script"; diff --git a/py/torch_tensorrt/_utils.py b/py/torch_tensorrt/_utils.py index f59dce9b1c..7dd1198cd7 100644 --- a/py/torch_tensorrt/_utils.py +++ b/py/torch_tensorrt/_utils.py @@ -55,10 +55,15 @@ def is_tensorrt_version_supported(min_version: str) -> bool: try: if trt._package_name == "tensorrt_rtx": return True - from importlib import metadata from packaging.version import Version + module = sys.modules["tensorrt"] + if module is not None and hasattr(module, "__version__"): + return bool(Version(module.__version__) >= Version(min_version)) + # if cannot get from the modules, fall back to metadata + from importlib import metadata + return bool(Version(metadata.version("tensorrt")) >= Version(min_version)) except (ImportError, ValueError): # If tensorrt is not installed or version cannot be determined diff --git a/tests/py/dynamo/conversion/test_pdist_aten.py b/tests/py/dynamo/conversion/test_pdist_aten.py index a7843780b8..d3b0fea5c4 100644 --- a/tests/py/dynamo/conversion/test_pdist_aten.py +++ b/tests/py/dynamo/conversion/test_pdist_aten.py @@ -64,13 +64,15 @@ class TestDynamicShapePdistConverter(DispatchTestCase): (5, 4), float("inf"), ), - ( - "dim0_dynamic_dim1_dynamic_p_other", - (2, 1), - (3, 2), - (4, 7), - 1.7, - ), + # disable this testcase due to https://github.com/pytorch/TensorRT/issues/3898 + # TODO: enable back once the issue is fixed in both rtx 1.2 and tensorrt 10.14 + # ( + # "dim0_dynamic_dim1_dynamic_p_other", + # (2, 1), + # (3, 2), + # (4, 7), + # 1.7, + # ), ] ) def test_pdist_float(self, _, min_shape, opt_shape, max_shape, p): From f91f167484b66ca94916869cddd9c21e51059a05 Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 23:02:51 -0800 Subject: [PATCH 13/14] modify --- .github/workflows/build-test-linux-x86_64_rtx.yml | 1 + .github/workflows/windows-test.yml | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/.github/workflows/build-test-linux-x86_64_rtx.yml b/.github/workflows/build-test-linux-x86_64_rtx.yml index c771b4ce67..4c1901ee7c 100644 --- a/.github/workflows/build-test-linux-x86_64_rtx.yml +++ b/.github/workflows/build-test-linux-x86_64_rtx.yml @@ -246,6 +246,7 @@ jobs: test-infra-ref: main build-matrix: ${{ needs.filter-matrix.outputs.matrix }} pre-script: ${{ matrix.pre-script }} + use-rtx: true script: | set -euo pipefail pushd . diff --git a/.github/workflows/windows-test.yml b/.github/workflows/windows-test.yml index 5e3b5c9918..e1d7e5916d 100644 --- a/.github/workflows/windows-test.yml +++ b/.github/workflows/windows-test.yml @@ -143,7 +143,7 @@ jobs: is_windows: 'enabled' - name: Surface failing tests if: always() - uses: pmeier/pytest-results-action@v0.3.0 + uses: pmeier/pytest-results-action@v0.7.2 with: path: ${{ env.RUNNER_TEST_RESULTS_DIR }} fail-on-empty: false From 9fbffc84a5f6cfcf52da47b7cf9c5716a337b58b Mon Sep 17 00:00:00 2001 From: lanluo-nvidia Date: Fri, 14 Nov 2025 23:42:22 -0800 Subject: [PATCH 14/14] modify --- .github/workflows/windows-test.yml | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/.github/workflows/windows-test.yml b/.github/workflows/windows-test.yml index e1d7e5916d..4753e31dcb 100644 --- a/.github/workflows/windows-test.yml +++ b/.github/workflows/windows-test.yml @@ -143,10 +143,12 @@ jobs: is_windows: 'enabled' - name: Surface failing tests if: always() - uses: pmeier/pytest-results-action@v0.7.2 + uses: pmeier/pytest-results-action@v0.5.0 with: - path: ${{ env.RUNNER_TEST_RESULTS_DIR }} - fail-on-empty: false + path: ${{ env.RUNNER_TEST_RESULTS_DIR }}/*.xml + summary: true + display-options: fEs + fail-on-empty: true - name: Teardown Windows if: ${{ always() }} uses: ./test-infra/.github/actions/teardown-windows