Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
50 commits
Select commit Hold shift + click to select a range
b9b7530
[webgpu] Update wgsl_templates README.md (#25336)
daijh Jul 9, 2025
bc0256f
[webgpu] Move the early return after copying for ScatterND (#25345)
qjia7 Jul 9, 2025
3b259e1
[EP ABI] Utility to serialize OrtGraph to GraphProto (#25292)
adrianlizarraga Jul 9, 2025
581bb20
Update vcpkg.json: remove optional-lite (#25339)
snnn Jul 9, 2025
a7178fd
Move buffer release or cache from OnRefresh to ReleaseBuffer in Bucke…
feich-ms Jul 9, 2025
e17ec57
[web] Fix "npm run pull:wasm" script (#25330)
fs-eire Jul 9, 2025
b49fc62
[MLAS] DequantizeLinear int8/uint8 (#24818)
adrianlizarraga Jul 9, 2025
cd5f91f
[CPU] GQA supports head_sink input for smooth softmax (#25269)
tianleiwu Jul 9, 2025
14e0ad7
Add PackageVersion parameter to NuGet packaging stage (#25315)
vraspar Jul 9, 2025
d293285
[QNN EP] Fix pool with reshape name conflicts (#25332)
qti-yuduo Jul 10, 2025
ff81567
Added creation of QDQ for TopK node (#25309)
Akupadhye Jul 10, 2025
d3820a2
[WebNN] Refactor webnn op input rank check and add validation for ops…
NingW101 Jul 10, 2025
8a27eab
Make TRT plugins optional (#25261)
gedoensmax Jul 10, 2025
e6658c0
[EP ABI] Add Graph_GetGraphView API to get a OrtGraph from a subset o…
chilo-ms Jul 10, 2025
591003b
[webgpu] a few optimization to WGSL template (#25333)
fs-eire Jul 10, 2025
57c9743
add --client_package_build option (#25351)
jywu-msft Jul 11, 2025
fb0f6c6
[WebNN] Fix bug in Float16Array availability check (#25354)
Honry Jul 11, 2025
9fc41c3
[EP ABI] Add Node_GetEpType API (#25350)
chilo-ms Jul 11, 2025
cee25ba
QNN-EP: DSPQueue Polling (#25361)
quic-calvnguy Jul 11, 2025
56078fe
[QNN_EP] Implement Efficient Mode API (#25146)
quic-calvnguy Jul 11, 2025
56a93a0
Add Compile API to set the location for the context binary file (#25356)
HectorSVC Jul 11, 2025
0d6e2d9
add build matrix for wgsl template (#25352)
fs-eire Jul 11, 2025
a532c8a
[JSEP] Fix inputShape index OOB in slice.ts (#25364)
jchen10 Jul 11, 2025
6ef13e3
[webgpu] extend cast version to 23 (#25235)
xhcao Jul 11, 2025
47b378d
Fix a security warning (#18979)
wschin Jul 11, 2025
a2b7a48
Fix AutoEpSelection and OrtEpLibrary tests when using AuthenticAMD (#…
quic-hungjuiw Jul 11, 2025
4ac95ce
Missing datatype in assertion (#23578)
ianfhunter Jul 11, 2025
aa644e8
[EP ABI] Update to use Node_GetEpName (#25363)
chilo-ms Jul 12, 2025
2b8c555
Bump clang-format from 20.1.7 to 20.1.8 (#25381)
dependabot[bot] Jul 14, 2025
613d22d
Fix number of layers in Whisper export (#25375)
kunal-vaishnavi Jul 14, 2025
491f435
Bump transformers from 4.48.0 to 4.52.1 in /onnxruntime/python/tools/…
dependabot[bot] Jul 14, 2025
10e2c1e
Bump ruff from 0.12.2 to 0.12.3 (#25382)
dependabot[bot] Jul 14, 2025
440ac68
[QNN EP] Upgrade QNN to 2.36.1 (#25388)
qti-jkilpatrick Jul 14, 2025
fab3069
Add vendor id to OrtEpFactory and default ORT logger to CreateEpFacto…
skottmckay Jul 15, 2025
9de58ac
Bump lintrunner-adapters from 0.12.4 to 0.12.5 (#25380)
dependabot[bot] Jul 15, 2025
d9ce6a9
[WebNN] Add rank range validation for rest ops (#25383)
Honry Jul 15, 2025
1e5fdd1
Fix some test issues when WebGPU and DML are enabled in the same buil…
skottmckay Jul 15, 2025
f19bb3c
Fix SigLIP casual mask bug (#25360)
nenad1002 Jul 15, 2025
c7250f4
[CPU] GQA supports attention scores output (#25319)
derdeljan-msft Jul 16, 2025
c7152ce
[QNN-EP] Support GridSample of linear mode for ONNX opset 20+ (#25408)
qti-yuduo Jul 16, 2025
7290edc
[QNN-EP] Update ScatterND op to reject only QNN-CPU (#25403)
qti-yuduo Jul 16, 2025
4a730ca
Fix 2 device discovery issues. (#25397)
skottmckay Jul 16, 2025
0955ab2
[webgpu] fix Slice implementation (#25415)
fs-eire Jul 16, 2025
d22a781
[QNN EP] Gpu backend test framework & test cases. (#25393)
johnpaultaken Jul 16, 2025
2d6a525
Fix Build Error when tensor dumping is enabled (#25414)
tianleiwu Jul 16, 2025
58954ba
[webgpu] Apply template to `MatMulNBitsWideTile` (#25353)
daijh Jul 16, 2025
5af86e5
Update docker images (#25418)
snnn Jul 16, 2025
3a5f75b
revert qnn sdk version (#25426)
prathikr Jul 17, 2025
7fe617c
Restore ability to handle non-hex string in device discovery vendor/d…
skottmckay Jul 17, 2025
bda52f6
Merge branch 'master' into sync_msft_17_7_25
ankitm3k Jul 17, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
40 changes: 0 additions & 40 deletions .github/workflows/linux-dnnl.yml

This file was deleted.

40 changes: 0 additions & 40 deletions .github/workflows/linux_migraphx_ci.yml

This file was deleted.

2 changes: 2 additions & 0 deletions .github/workflows/windows_webgpu.yml
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ jobs:
strategy:
matrix:
vcpkg_option: [novcpkg, vcpkg]
wgsl_template: [static, dynamic]
env:
OrtPackageId: Microsoft.ML.OnnxRuntime
OnnxRuntimeBuildDirectory: ${{ github.workspace }}
Expand Down Expand Up @@ -123,6 +124,7 @@ jobs:
--build_nodejs `
--build_java `
--use_webgpu `
--wgsl_template ${{ matrix.wgsl_template }} `
${{ matrix.vcpkg_option == 'vcpkg' && '--use_vcpkg' || '' }} `
--cmake_extra_defines `
onnxruntime_BUILD_UNIT_TESTS=ON `
Expand Down
132 changes: 0 additions & 132 deletions .github/workflows/windows_x64_release_dnnl_build_x64_release.yml

This file was deleted.

1 change: 1 addition & 0 deletions cmake/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -151,6 +151,7 @@ option(onnxruntime_DISABLE_SPARSE_TENSORS "Disable sparse tensors data types" OF
option(onnxruntime_DISABLE_OPTIONAL_TYPE "Disable optional type" OFF)
option(onnxruntime_DISABLE_FLOAT8_TYPES "Disable float 8 types" OFF)
option(onnxruntime_MINIMAL_BUILD "Exclude as much as possible from the build. Support ORT format models. No support for ONNX format models." OFF)
option(onnxruntime_CLIENT_PACKAGE_BUILD "Enables default settings that are more appropriate for client/on-device workloads." OFF)
cmake_dependent_option(onnxruntime_DISABLE_RTTI "Disable RTTI" ON "NOT onnxruntime_ENABLE_PYTHON;NOT onnxruntime_USE_CUDA" OFF)
# For now onnxruntime_DISABLE_EXCEPTIONS will only work with onnxruntime_MINIMAL_BUILD, more changes (ONNX, non-CPU EP, ...) are required to run this standalone
cmake_dependent_option(onnxruntime_DISABLE_EXCEPTIONS "Disable exception handling. Requires onnxruntime_MINIMAL_BUILD currently." ON "onnxruntime_MINIMAL_BUILD;NOT onnxruntime_ENABLE_PYTHON" OFF)
Expand Down
5 changes: 5 additions & 0 deletions cmake/adjust_global_compile_flags.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -95,6 +95,11 @@ if (onnxruntime_MINIMAL_BUILD)
endif()
endif()

# ORT build with default settings more appropriate for client/on-device workloads.
if (onnxruntime_CLIENT_PACKAGE_BUILD)
add_compile_definitions(ORT_CLIENT_PACKAGE_BUILD)
endif()

if (onnxruntime_ENABLE_LTO)
include(CheckIPOSupported)
check_ipo_supported(RESULT ipo_enabled OUTPUT ipo_output)
Expand Down
25 changes: 18 additions & 7 deletions cmake/external/onnxruntime_external_deps.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -774,13 +774,24 @@ if (onnxruntime_USE_WEBGPU)
endif()

if (NOT CMAKE_SYSTEM_NAME STREQUAL "Emscripten" AND onnxruntime_WGSL_TEMPLATE STREQUAL "dynamic")
onnxruntime_fetchcontent_declare(
duktape
URL ${DEP_URL_duktape}
URL_HASH SHA1=${DEP_SHA1_duktape}
EXCLUDE_FROM_ALL
)
onnxruntime_fetchcontent_makeavailable(duktape)
if(onnxruntime_USE_VCPKG)
find_package(unofficial-duktape CONFIG REQUIRED)
add_library(duktape_static ALIAS unofficial::duktape::duktape)
else()
onnxruntime_fetchcontent_declare(
duktape
URL ${DEP_URL_duktape}
URL_HASH SHA1=${DEP_SHA1_duktape}
EXCLUDE_FROM_ALL
)
onnxruntime_fetchcontent_makeavailable(duktape)

if(NOT TARGET duktape_static)
add_library(duktape_static STATIC "${duktape_SOURCE_DIR}/src/duktape.c")
target_compile_features(duktape_static PRIVATE c_std_99)
target_include_directories(duktape_static INTERFACE $<BUILD_INTERFACE:${duktape_SOURCE_DIR}/src>)
endif()
endif()
endif()
endif()

Expand Down
1 change: 1 addition & 0 deletions cmake/onnxruntime_mlas.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,7 @@ onnxruntime_add_static_library(onnxruntime_mlas
${MLAS_SRC_DIR}/eltwise.cpp
${MLAS_SRC_DIR}/erf.cpp
${MLAS_SRC_DIR}/compute.cpp
${MLAS_SRC_DIR}/dequantize.cpp
${MLAS_SRC_DIR}/quantize.cpp
${MLAS_SRC_DIR}/qgemm_kernel_default.cpp
${MLAS_SRC_DIR}/qladd.cpp
Expand Down
23 changes: 5 additions & 18 deletions cmake/onnxruntime_providers_tensorrt.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -72,26 +72,21 @@
endif()

# TensorRT 10 GA onwards, the TensorRT libraries will have major version appended to the end on Windows,
# for example, nvinfer_10.dll, nvinfer_plugin_10.dll, nvonnxparser_10.dll ...
# for example, nvinfer_10.dll, nvonnxparser_10.dll ...
if (WIN32 AND TRT_GREATER_OR_EQUAL_TRT_10_GA)
set(NVINFER_LIB "nvinfer_${NV_TENSORRT_MAJOR}")
set(NVINFER_PLUGIN_LIB "nvinfer_plugin_${NV_TENSORRT_MAJOR}")
set(PARSER_LIB "nvonnxparser_${NV_TENSORRT_MAJOR}")
endif()

if (NOT NVINFER_LIB)
set(NVINFER_LIB "nvinfer")
endif()

if (NOT NVINFER_PLUGIN_LIB)
set(NVINFER_PLUGIN_LIB "nvinfer_plugin")
endif()

if (NOT PARSER_LIB)
set(PARSER_LIB "nvonnxparser")
endif()

MESSAGE(STATUS "Looking for ${NVINFER_LIB} and ${NVINFER_PLUGIN_LIB}")
MESSAGE(STATUS "Looking for ${NVINFER_LIB}")

find_library(TENSORRT_LIBRARY_INFER ${NVINFER_LIB}
HINTS ${TENSORRT_ROOT}
Expand All @@ -101,14 +96,6 @@
MESSAGE(STATUS "Can't find ${NVINFER_LIB}")
endif()

find_library(TENSORRT_LIBRARY_INFER_PLUGIN ${NVINFER_PLUGIN_LIB}
HINTS ${TENSORRT_ROOT}
PATH_SUFFIXES lib lib64 lib/x64)

if (NOT TENSORRT_LIBRARY_INFER_PLUGIN)
MESSAGE(STATUS "Can't find ${NVINFER_PLUGIN_LIB}")
endif()

if (onnxruntime_USE_TENSORRT_BUILTIN_PARSER)
MESSAGE(STATUS "Looking for ${PARSER_LIB}")

Expand All @@ -120,7 +107,7 @@
MESSAGE(STATUS "Can't find ${PARSER_LIB}")
endif()

set(TENSORRT_LIBRARY ${TENSORRT_LIBRARY_INFER} ${TENSORRT_LIBRARY_INFER_PLUGIN} ${TENSORRT_LIBRARY_NVONNXPARSER})
set(TENSORRT_LIBRARY ${TENSORRT_LIBRARY_INFER} ${TENSORRT_LIBRARY_NVONNXPARSER})
MESSAGE(STATUS "Find TensorRT libs at ${TENSORRT_LIBRARY}")
else()
if (TRT_GREATER_OR_EQUAL_TRT_10_GA)
Expand Down Expand Up @@ -153,15 +140,15 @@
endif()
# Static libraries are just nvonnxparser_static on all platforms
set(onnxparser_link_libs nvonnxparser_static)
set(TENSORRT_LIBRARY ${TENSORRT_LIBRARY_INFER} ${TENSORRT_LIBRARY_INFER_PLUGIN})
set(TENSORRT_LIBRARY ${TENSORRT_LIBRARY_INFER})
MESSAGE(STATUS "Find TensorRT libs at ${TENSORRT_LIBRARY}")
endif()

# ${TENSORRT_LIBRARY} is empty if we link nvonnxparser_static.
# nvonnxparser_static is linked against tensorrt libraries in onnx-tensorrt
# See https://github.com/onnx/onnx-tensorrt/blob/8af13d1b106f58df1e98945a5e7c851ddb5f0791/CMakeLists.txt#L121
# However, starting from TRT 10 GA, nvonnxparser_static doesn't link against tensorrt libraries.
# Therefore, the above code finds ${TENSORRT_LIBRARY_INFER} and ${TENSORRT_LIBRARY_INFER_PLUGIN}.
# Therefore, the above code finds ${TENSORRT_LIBRARY_INFER}.
if(onnxruntime_CUDA_MINIMAL)
set(trt_link_libs ${CMAKE_DL_LIBS} ${TENSORRT_LIBRARY})
else()
Expand Down
Loading
Loading