Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
62 commits
Select commit Hold shift + click to select a range
e6fc589
[QNN EP] Transform SFIXED to UFIXED on InstanceNorm scale (#24952)
quic-hungjuiw Jun 11, 2025
0e027fe
[webgpu] Bump version of Dawn to 991810463a (#25011)
fs-eire Jun 11, 2025
ce503d7
Upgrade QNN to 2.35.0 (#25002)
qti-jkilpatrick Jun 11, 2025
0e4bf97
Fix illegal update model opset version in static_quantize_runner (#24…
zhaoxul-qti Jun 11, 2025
9f0d989
Add Whisper JSON files for ONNX Runtime GenAI (#25032)
kunal-vaishnavi Jun 11, 2025
7c7b6e0
[QNN EP] Enable boolean support for expand op (#25016)
qti-yuduo Jun 11, 2025
11f0a0a
Convert graph initializers into OrtValue Phase I (#23979)
yuslepukhin Jun 11, 2025
2ea4ea4
Fix how Whisper tokenizer is loaded and saved (#25037)
kunal-vaishnavi Jun 12, 2025
70f14d7
[CUDA] Suppress MSVC build warnings for sm=90 (#25020)
tianleiwu Jun 12, 2025
a643acf
[QNN EP] Add fixing dynamic input shapes in qnn.preprocess (#25024)
huaychou Jun 12, 2025
2b43536
[test] add retry to onnx node test data download (#25034)
fs-eire Jun 12, 2025
30c5f05
Add Paged Attention Op for CUDA SM80 support (#24595)
aciddelgado Jun 12, 2025
3b855e1
Support 8 bit weights "unpacked" compute mode in MatmulNBits kernel (…
hariharans29 Jun 12, 2025
d8096f4
[WebNN] Fix bug in getting the first input name of a WebNN op (#25027)
Honry Jun 12, 2025
49e7717
Refactor the cmake code of onnxruntime_shared_lib_dlopen_test (#25031)
snnn Jun 12, 2025
2c29984
[WebNN] Expose Module.webnnRegisterMLContext to EM_ASM block (#25022)
Honry Jun 12, 2025
ab61dc4
Add paralell to win-ci.yml (#25039)
snnn Jun 12, 2025
a5494ed
Make OrtDevice generic to support plugin EP usage. (#24985)
skottmckay Jun 13, 2025
9ffc650
[WebGPU-EP] Disable SubgroupMatrix uniformity checks (#25028)
jchen10 Jun 13, 2025
82fddd7
Cherry pick telemetry changes from win-onnxruntime (#24957)
ashrit-ms Jun 13, 2025
0784e0a
Revert "Improve Windows ETW callback registration and fix issues" (#2…
tianleiwu Jun 13, 2025
e7c9a6c
[webgpu] make DP4AMatMulNBitsSmallMProgram shader template (#25025)
jing-bao Jun 13, 2025
b3c316f
[CUDA] Fix clip operator (#25057)
tianleiwu Jun 14, 2025
82c1bf9
[webgpu] support intel subgroup matrix on matmul_nbits (#24898)
xhcao Jun 14, 2025
ece0c1e
Disable RMSNorm in onnx_backend_test.py and Topk(uint_64) (#25056)
titaiwangms Jun 16, 2025
089c52e
Add python bindings to the global thread pool functionality (#24238)
khoover Jun 16, 2025
2594b47
Fix some `OrtStatus*` leaks (#25036)
edgchen1 Jun 16, 2025
53f3c6c
[WebGPU EP] adjusts pow op to use sqrt when exponent is 0.5 (#25026)
prathikr Jun 16, 2025
7572970
Fix CANN build error (#25075)
bachelor-dou Jun 16, 2025
2df3983
Skip topk uint64 test added from ONNX 1.18 (#25084)
titaiwangms Jun 16, 2025
20b2247
vitisai/imp/graph.cc: fix graph_save() for models with external data …
schuermans-slx Jun 16, 2025
796ad45
[webgpu] Add zero points support for dp4 path (#24675)
qjia7 Jun 17, 2025
7695233
Add version check for IAlloactor::GetStats (#24989)
toothache Jun 17, 2025
0699df6
[webgpu] Bump version of Dawn to 9733be39e (#25071)
fs-eire Jun 17, 2025
3eeff82
[WebGPU EP] fixes bugs in NCHW version of instance norm operator (#25…
prathikr Jun 18, 2025
13c0631
fix webgpu linux build on gcc 13.3 (#25099)
guschmue Jun 18, 2025
0385779
[QNN EP] Add Support for Reciprocal Op in QNN EP (#25035)
quic-ashwshan Jun 19, 2025
0ef8213
Add GetCapability/Compile infrastructure for EP ABI (#24887)
adrianlizarraga Jun 19, 2025
6413908
Skip hash computation for EPContext models (#25106)
ashrit-ms Jun 19, 2025
74126d1
Fix C/C++ documentation generation (#25112)
adrianlizarraga Jun 19, 2025
dbfbebe
[WebNN] Handle in-memory external data (#25079)
Honry Jun 19, 2025
2c041e0
use WebGPU EP instead of JSEP in WebAssembly (#24856)
fs-eire Jun 19, 2025
bc95e99
Update abseil (#25033)
snnn Jun 20, 2025
e0a4ed1
Use model hash from metadata if available (#25118)
ashrit-ms Jun 20, 2025
02aa881
[wasm] Optimize WASM relaxed simd MlasGemmQuantKernel (#25048)
fanchenkong1 Jun 20, 2025
7268117
[CUDA] FpA IntB Gemm Kernel Test (#25109)
tianleiwu Jun 20, 2025
ae628b9
Fix: delay CUDADriverWrapper instantiation to avoid uncaught exceptio…
krahenbuhl Jun 20, 2025
c98a0e0
update CMAKE_CUDA_ARCHITECTURES for packaging pipelines (#25126)
tianleiwu Jun 21, 2025
b7b1af4
[QNN EP] Add remove_initializer_from_input in qnn.preprocess (#25102)
quic-hungjuiw Jun 23, 2025
a0c4236
[NV TensorRT RTX EP] enable weight stripped engines with EP Context (…
thevishalagarwal Jun 23, 2025
be0292f
[jsep-webgpu] Add kernel profiling start time in logging (#25132)
daijh Jun 23, 2025
ab4d85c
[TRT RTX EP] fix build error (#25094)
thevishalagarwal Jun 23, 2025
986b66a
[QNN EP] Add QNN EP UDO support (#25076)
chenweng-quic Jun 23, 2025
15f6bde
[webgpu] Make the GQA's intermediate buffer static (#25091)
qjia7 Jun 24, 2025
a96e7c9
[QNN-EP] Fix the input order in gen_qnn_ctx_onn_model.py (#25090)
chuteng-quic Jun 24, 2025
4e45b29
[wasm] Optimize WASM SIMD MlasGemmQuantKernel (#25136)
fanchenkong1 Jun 24, 2025
3a47bd2
Fix default value for Whisper's buffer sharing (#25151)
kunal-vaishnavi Jun 24, 2025
7402b6c
Enable warnings in ONNX build through onnx.patch and vcpkg binskim.pa…
titaiwangms Jun 24, 2025
6cffd1a
Add RotaryEmbeddings(23) - CPU (#24980)
titaiwangms Jun 25, 2025
f80e6f4
Enable VTCM Back Up Buffer Sharing (#24962)
quic-calvnguy Jun 25, 2025
7d22c09
[EP ABI] Add support for creating EP Context models. (#25124)
adrianlizarraga Jun 25, 2025
7176c12
Merge branch 'master' into sync_msft_25_6_25
ankitm3k Jun 25, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
5 changes: 2 additions & 3 deletions .github/workflows/linux-wasm-ci-build-and-test-workflow.yml
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,6 @@ jobs:
${{ env.common_build_args }} \
--build_dir ${{ github.workspace }}/build/wasm_inferencing_webgpu \
--use_webgpu \
--use_jsep \
--use_webnn \
--target onnxruntime_webassembly \
--skip_tests
Expand All @@ -113,8 +112,8 @@ jobs:
if: ${{ inputs.skip_publish != true && inputs.build_webgpu == true }}
run: |
mkdir -p ${{ github.workspace }}/artifacts/wasm_webgpu/
cp ${{ github.workspace }}/build/wasm_inferencing_webgpu/${{ inputs.build_config }}/ort-wasm-simd-threaded.jsep.wasm ${{ github.workspace }}/artifacts/wasm_webgpu/
cp ${{ github.workspace }}/build/wasm_inferencing_webgpu/${{ inputs.build_config }}/ort-wasm-simd-threaded.jsep.mjs ${{ github.workspace }}/artifacts/wasm_webgpu/
cp ${{ github.workspace }}/build/wasm_inferencing_webgpu/${{ inputs.build_config }}/ort-wasm-simd-threaded.asyncify.wasm ${{ github.workspace }}/artifacts/wasm_webgpu/
cp ${{ github.workspace }}/build/wasm_inferencing_webgpu/${{ inputs.build_config }}/ort-wasm-simd-threaded.asyncify.mjs ${{ github.workspace }}/artifacts/wasm_webgpu/

- name: Upload WASM artifacts
if: ${{ inputs.skip_publish != true }}
Expand Down
55 changes: 18 additions & 37 deletions .github/workflows/windows-web-ci-workflow.yml
Original file line number Diff line number Diff line change
Expand Up @@ -16,9 +16,6 @@ on:
package_name:
type: string
default: "NPM_packages"
run_webgpu_tests:
type: boolean
default: true

jobs:
build_onnxruntime_web:
Expand Down Expand Up @@ -86,6 +83,22 @@ jobs:
run: |
copy ${{ github.workspace }}\artifacts_wasm\ort-*.mjs ${{ github.workspace }}\js\web\dist\

- name: Download WebAssembly WebGPU artifacts
uses: actions/download-artifact@v4
with:
name: ${{ inputs.build_config }}_wasm_webgpu
path: ${{ github.workspace }}/artifacts_wasm_webgpu

- name: Binplace dist files (.wasm) for WebGPU
shell: cmd
run: |
copy ${{ github.workspace }}\artifacts_wasm_webgpu\ort-*.wasm ${{ github.workspace }}\js\web\dist\

- name: Binplace dist files (.mjs) for WebGPU
shell: cmd
run: |
copy ${{ github.workspace }}\artifacts_wasm_webgpu\ort-*.mjs ${{ github.workspace }}\js\web\dist\

- name: npm ci for /js/
run: npm ci
working-directory: ${{ github.workspace }}/js
Expand Down Expand Up @@ -115,17 +128,7 @@ jobs:
run: |
Get-WmiObject Win32_Process -Filter "name = 'chrome.exe'" | Format-List CommandLine

- name: Run ort-web tests (wasm,webgl backend)
if: ${{ inputs.run_webgpu_tests != true }}
shell: cmd
run: |
mkdir ${{ runner.temp }}\web\test\01
dir ${{ runner.temp }}\web\test\01
npm test -- -e=chrome -b=webgl,wasm --user-data-dir=${{ runner.temp }}\web\test\01 --chromium-flags=--enable-logging --chromium-flags=--v=1
working-directory: ${{ github.workspace }}\js\web

- name: Run ort-web tests (ALL backends)
if: ${{ inputs.run_webgpu_tests == true }}
shell: cmd
run: |
mkdir ${{ runner.temp }}\web\test\02
Expand All @@ -134,7 +137,6 @@ jobs:
working-directory: ${{ github.workspace }}\js\web

- name: Run ort-web tests (Suite1, webgpu, IO-binding=gpu-tensor)
if: ${{ inputs.run_webgpu_tests == true }}
shell: cmd
run: |
mkdir ${{ runner.temp }}\web\test\03
Expand All @@ -143,7 +145,6 @@ jobs:
working-directory: ${{ github.workspace }}\js\web

- name: Run ort-web tests (Suite1, webgpu, IO-binding=gpu-location)
if: ${{ inputs.run_webgpu_tests == true }}
shell: cmd
run: |
mkdir ${{ runner.temp }}\web\test\04
Expand All @@ -169,27 +170,7 @@ jobs:
working-directory: ${{ github.workspace }}\js\web

# WebGPU EP tests
- name: Download WebAssembly WebGPU artifacts
if: ${{ inputs.run_webgpu_tests == true }}
uses: actions/download-artifact@v4
with:
name: ${{ inputs.build_config }}_wasm_webgpu
path: ${{ github.workspace }}/artifacts_wasm_webgpu

- name: Binplace dist files (.wasm) for WebGPU
if: ${{ inputs.run_webgpu_tests == true }}
shell: cmd
run: |
copy /Y ${{ github.workspace }}\artifacts_wasm_webgpu\ort-*.wasm ${{ github.workspace }}\js\web\dist\

- name: Binplace dist files (.mjs) for WebGPU
if: ${{ inputs.run_webgpu_tests == true }}
shell: cmd
run: |
copy /Y ${{ github.workspace }}\artifacts_wasm_webgpu\ort-*.mjs ${{ github.workspace }}\js\web\dist\

- name: Run ort-web tests - WebGPU EP
if: ${{ inputs.run_webgpu_tests == true }}
continue-on-error: true
shell: cmd
run: |
Expand All @@ -199,15 +180,15 @@ jobs:
working-directory: ${{ github.workspace }}\js\web

- name: Validate shader keys - WebGPU EP
if: ${{ inputs.run_webgpu_tests == true && inputs.build_config == 'Debug' }}
if: ${{ inputs.build_config == 'Debug' }}
uses: ./.github/actions/webgpu-validate-shader-key
with:
log_file_path: ${{ runner.temp }}\web\test\07\chrome_debug.log
is_chromium_log: true

# this step is added to help investigate the shader validation failure which is hard to reproduce
- name: Upload WebGPU shader validation log on failure
if: ${{ failure() && inputs.run_webgpu_tests == true && inputs.build_config == 'Debug' }}
if: ${{ failure() && inputs.build_config == 'Debug' }}
uses: actions/upload-artifact@v4
with:
name: webgpu-shader-validation-logs
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/windows_x86.yml
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ jobs:
- name: Build and Test
shell: pwsh
run: |
python.exe "${{ github.workspace }}\tools\ci_build\build.py" --config RelWithDebInfo --build_dir "${{ github.workspace }}\build" --skip_submodule_sync --build_csharp --parallel --use_binskim_compliant_compile_flags --cmake_generator "Visual Studio 17 2022" --build_shared_lib --enable_onnx_tests --build_wheel --msbuild_extra_options "IncludeMobileTargets=false" --build_nuget --use_vcpkg --use_vcpkg_ms_internal_asset_cache
python.exe "${{ github.workspace }}\tools\ci_build\build.py" --config RelWithDebInfo --build_dir "${{ github.workspace }}\build" --skip_submodule_sync --build_csharp --parallel --use_binskim_compliant_compile_flags --cmake_generator "Visual Studio 17 2022" --build_shared_lib --enable_onnx_tests --build_wheel --msbuild_extra_options "IncludeMobileTargets=false" --build_nuget --compile_no_warning_as_error --use_vcpkg --use_vcpkg_ms_internal_asset_cache
if ($LASTEXITCODE -ne 0) {
exit $LASTEXITCODE
}
Expand Down
8 changes: 6 additions & 2 deletions cmake/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -372,7 +372,7 @@ if (onnxruntime_USE_ROCM)
if (HIPIFY_PERL_PATH-NOTFOUND)
MESSAGE(FATAL_ERROR "hipify-perl not found")
endif()
MESSAGE("HIPIFY PATH:"${HIPIFY_PERL_PATH}/hipify-perl)
MESSAGE("HIPIFY PATH: ${HIPIFY_PERL_PATH}/hipify-perl")
set(onnxruntime_HIPIFY_PERL ${HIPIFY_PERL_PATH}/hipify-perl)
endif()

Expand Down Expand Up @@ -1336,7 +1336,7 @@ function(onnxruntime_configure_target target_name)
if(WIN32 AND onnxruntime_ENABLE_STATIC_ANALYSIS AND onnxruntime_USE_CUSTOM_STATIC_ANALYSIS_RULES)
set_target_properties(${target_name} PROPERTIES VS_USER_PROPS ${PROJECT_SOURCE_DIR}/EnableVisualStudioCodeAnalysis.props)
endif()
target_include_directories(${target_name} PRIVATE ${CMAKE_CURRENT_BINARY_DIR} ${ONNXRUNTIME_ROOT} ${abseil_cpp_SOURCE_DIR})
target_include_directories(${target_name} PRIVATE ${CMAKE_CURRENT_BINARY_DIR} ${ONNXRUNTIME_ROOT})
if (onnxruntime_ENABLE_TRAINING_OPS)
target_include_directories(${target_name} PRIVATE ${ORTTRAINING_ROOT})
endif()
Expand Down Expand Up @@ -1669,6 +1669,10 @@ if (onnxruntime_ENABLE_DLPACK)
add_compile_definitions(ENABLE_DLPACK)
endif()

if (onnxruntime_CALLER_FRAMEWORK)
add_definitions(-DORT_CALLER_FRAMEWORK="${onnxruntime_CALLER_FRAMEWORK}")
endif()

if (UNIX OR onnxruntime_USE_NCCL)
# Find NCCL
if (onnxruntime_USE_NCCL)
Expand Down
4 changes: 2 additions & 2 deletions cmake/deps.txt
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
#since the file contains a version string: "lts_20230802". However, the file is for debugging purposes only and would
#not affect built binaries.
#
abseil_cpp;https://github.com/abseil/abseil-cpp/archive/refs/tags/20240722.0.zip;36ee53eb1466fb6e593fc5c286680de31f8a494a
abseil_cpp;https://github.com/abseil/abseil-cpp/archive/refs/tags/20250512.0.zip;3d6ff7e7ce144d9a53a53bef1f1bf79e1da4b8e1
coremltools;https://github.com/apple/coremltools/archive/refs/tags/7.1.zip;f1bab0f30966f2e217d8e01207d518f230a1641a
cxxopts;https://github.com/jarro2783/cxxopts/archive/3c73d91c0b04e2b59462f0a741be8c07024c1bc0.zip;6c6ca7f8480b26c8d00476e0e24b7184717fe4f0
date;https://github.com/HowardHinnant/date/archive/refs/tags/v3.0.1.zip;2dac0c81dc54ebdd8f8d073a75c053b04b56e159
Expand Down Expand Up @@ -56,5 +56,5 @@ extensions;https://github.com/microsoft/onnxruntime-extensions/archive/c24b7bab0
composable_kernel;https://github.com/ROCmSoftwarePlatform/composable_kernel/archive/204da9c522cebec5220bba52cd3542ebcaf99e7a.zip;1827348efd47831c13074245274d41b7cae8a557
directx_headers;https://github.com/microsoft/DirectX-Headers/archive/refs/tags/v1.613.1.zip;47653509a3371eabb156360f42faf582f314bf2e
cudnn_frontend;https://github.com/NVIDIA/cudnn-frontend/archive/refs/tags/v1.12.0.zip;7e733cfdc410d777b76122d64232499205589a96
dawn;https://github.com/google/dawn/archive/4cb1f9be152a4fa6bb695c08cd707ab078a1e2fb.zip;de39336b7715f53c14eec61072293b85cc73b691
dawn;https://github.com/google/dawn/archive/9733be39e18186961d503e064874afe3e9ceb8d1.zip;2a4017c32892b90d072a9102eba90ae691fae36d
kleidiai;https://github.com/ARM-software/kleidiai/archive/refs/tags/v1.4.0.tar.gz;22d3b57b54a61c194ab256ff11b0353a3b220244
4 changes: 2 additions & 2 deletions cmake/external/abseil-cpp.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ else()
endif()

# NB! Advancing Abseil version changes its internal namespace,
# currently absl::lts_20240116 which affects abseil-cpp.natvis debugger
# currently absl::lts_20250512 which affects abseil-cpp.natvis debugger
# visualization file, that must be adjusted accordingly, unless we eliminate
# that namespace at build time.
onnxruntime_fetchcontent_declare(
Expand All @@ -36,7 +36,7 @@ onnxruntime_fetchcontent_declare(
URL_HASH SHA1=${DEP_SHA1_abseil_cpp}
EXCLUDE_FROM_ALL
PATCH_COMMAND ${ABSL_PATCH_COMMAND}
FIND_PACKAGE_ARGS 20240722 NAMES absl
FIND_PACKAGE_ARGS 20250512 NAMES absl
)

onnxruntime_fetchcontent_makeavailable(abseil_cpp)
Expand Down
10 changes: 5 additions & 5 deletions cmake/external/abseil-cpp.natvis
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
<?xml version="1.0" encoding="utf-8"?>
<AutoVisualizer xmlns="http://schemas.microsoft.com/vstudio/debugger/natvis/2010">
<Type Name="absl::lts_20240722::InlinedVector&lt;*&gt;">
<Type Name="absl::lts_20250512::InlinedVector&lt;*&gt;">
<Intrinsic Name="_size" Expression="storage_.metadata_.value >> 1"/>
<Intrinsic Name="_is_allocated" Expression="(storage_.metadata_.value &amp; 1) == 1"/>
<Intrinsic Name="_inlined_data" Expression="($T1*)storage_.data_.inlined.inlined_data"/>
Expand All @@ -24,7 +24,7 @@
</Expand>
</Type>
<!-- Should handle both flat hash_set and hash_map -->
<Type Name="absl::lts_20240116::container_internal::raw_hash_set&lt;*&gt;">
<Type Name="absl::lts_20250512::container_internal::raw_hash_set&lt;*&gt;">
<Intrinsic Name="_commonfields" Expression="settings_.value"/>
<Intrinsic Name="_size" Expression="settings_.value.compressed_tuple_.value"/>
<Intrinsic Name="_capacity" Expression="_commonfields().capacity_"/>
Expand All @@ -51,7 +51,7 @@
</Type>

<!-- Primitive types stored as a value -->
<Type Name="absl::lts_20240116::container_internal::Storage&lt;*,*,0&gt;">
<Type Name="absl::lts_20250512::container_internal::Storage&lt;*,*,0&gt;">
<DisplayString IncludeView="noparens">*($T1 *){value}</DisplayString>
<DisplayString ExcludeView="noparens">(*($T1 *){value})</DisplayString>
<Expand>
Expand All @@ -60,15 +60,15 @@
</Type>

<!-- For storage inherited from the type -->
<Type Name="absl::lts_20240116::container_internal::Storage&lt;*,*,1&gt;">
<Type Name="absl::lts_20250512::container_internal::Storage&lt;*,*,1&gt;">
<DisplayString IncludeView="noparens">*($T1 *)this</DisplayString>
<DisplayString ExcludeView="noparens">(*($T1 *)this)</DisplayString>
<Expand>
<ExpandedItem>*($T1 *)this</ExpandedItem>
</Expand>
</Type>

<Type Name="absl::lts_20240116::container_internal::map_slot_type&lt;*&gt;">
<Type Name="absl::lts_20250512::container_internal::map_slot_type&lt;*&gt;">
<DisplayString IncludeView="noparens">{value.first}, {value.second}</DisplayString>
<DisplayString ExcludeView="noparens">({value.first}, {value.second})</DisplayString>
<Expand>
Expand Down
20 changes: 3 additions & 17 deletions cmake/external/onnxruntime_external_deps.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -723,36 +723,22 @@ if (onnxruntime_USE_WEBGPU)
)
else()
set(ONNXRUNTIME_Dawn_PATCH_COMMAND
# The dawn.patch contains the following changes:
# The dawn_destroy_buffer_on_destructor.patch contains the following changes:
#
# - (private) Allow WGPUBufferImpl class to destroy the buffer in the destructor
# In native implementation, wgpuBufferRelease will trigger the buffer destroy (if refcount decreased to 0). But
# in emwgpu implementation, the buffer destroy won't happen. This change adds a destructor to the buffer class
# to destroy the buffer when the refcount is 0 for non-external buffers.
#
# - (private) Remove hard-coded CMAKE_OSX_DEPLOYMENT_TARGET in Dawn's CMake files
# https://github.com/microsoft/onnxruntime/pull/23729
#
# - (private) Reduce unsafe buffer usage warning in aligned_storage.h
# https://github.com/microsoft/onnxruntime/pull/24308
# The patch disables the UNSAFE_BUFFER_USAGE warning around the AlignedStorage struct in aligned_storage.h. This is done
# by using TINT_BEGIN_DISABLE_WARNING and TINT_END_DISABLE_WARNING macros, which helps in warnings related to unsafe buffer usage
# usage when compiling the code, making the build process cleaner and faster.
#
${Patch_EXECUTABLE} --binary --ignore-whitespace -p1 < ${PROJECT_SOURCE_DIR}/patches/dawn/dawn.patch &&
${Patch_EXECUTABLE} --binary --ignore-whitespace -p1 < ${PROJECT_SOURCE_DIR}/patches/dawn/dawn_destroy_buffer_on_destructor.patch &&

# The dawn_force_enable_f16_nvidia_vulkan.patch contains the following changes:
#
# - (private) Force enable f16 support for NVIDIA Vulkan
# Dawn disabled f16 support for NVIDIA Vulkan by default because of crashes in f16 CTS tests (crbug.com/tint/2164).
# Since the crashes are limited to specific GPU models, we patched Dawn to remove the restriction.
${Patch_EXECUTABLE} --binary --ignore-whitespace -p1 < ${PROJECT_SOURCE_DIR}/patches/dawn/dawn_force_enable_f16_nvidia_vulkan.patch &&

# The dawn_fix_copy_dxil_dll.patch contains the following changes:
#
# - (private) Fix copy of dxil.dll in Dawn
# The patch ensures the copy of dxil.dll to be done after the build step of `dxcompiler` target.
${Patch_EXECUTABLE} --binary --ignore-whitespace -p1 < ${PROJECT_SOURCE_DIR}/patches/dawn/dawn_fix_copy_dxil_dll.patch)
${Patch_EXECUTABLE} --binary --ignore-whitespace -p1 < ${PROJECT_SOURCE_DIR}/patches/dawn/dawn_force_enable_f16_nvidia_vulkan.patch)

onnxruntime_fetchcontent_declare(
dawn
Expand Down
1 change: 1 addition & 0 deletions cmake/onnxruntime.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@ endif()
function(get_c_cxx_api_headers HEADERS_VAR)
set(_headers
"${REPO_ROOT}/include/onnxruntime/core/session/onnxruntime_c_api.h"
"${REPO_ROOT}/include/onnxruntime/core/session/onnxruntime_ep_c_api.h"
"${REPO_ROOT}/include/onnxruntime/core/session/onnxruntime_cxx_api.h"
"${REPO_ROOT}/include/onnxruntime/core/session/onnxruntime_cxx_inline.h"
"${REPO_ROOT}/include/onnxruntime/core/session/onnxruntime_float16.h"
Expand Down
14 changes: 6 additions & 8 deletions cmake/onnxruntime_common.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -120,16 +120,14 @@ if (onnxruntime_USE_MIMALLOC)
target_link_libraries(onnxruntime_common PRIVATE onnxruntime_mimalloc_shim)
endif()

if(NOT onnxruntime_DISABLE_ABSEIL)
target_include_directories(onnxruntime_common PRIVATE ${ABSEIL_SOURCE_DIR})
if (MSVC)
set(ABSEIL_NATVIS_FILE "abseil-cpp.natvis")
target_sources(
onnxruntime_common
INTERFACE $<BUILD_INTERFACE:${PROJECT_SOURCE_DIR}/external/${ABSEIL_NATVIS_FILE}>)
endif()
if (MSVC)
set(ABSEIL_NATVIS_FILE "abseil-cpp.natvis")
target_sources(
onnxruntime_common
INTERFACE $<BUILD_INTERFACE:${PROJECT_SOURCE_DIR}/external/${ABSEIL_NATVIS_FILE}>)
endif()


if (MSVC)
set(EIGEN_NATVIS_FILE ${eigen_SOURCE_DIR}/debug/msvc/eigen.natvis)
if (EXISTS ${EIGEN_NATVIS_FILE})
Expand Down
3 changes: 1 addition & 2 deletions cmake/onnxruntime_lora.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -10,8 +10,7 @@ file(GLOB onnxruntime_lora_srcs CONFIGURE_DEPENDS
source_group(TREE ${REPO_ROOT} FILES ${onnxruntime_lora_srcs})

onnxruntime_add_static_library(onnxruntime_lora ${onnxruntime_lora_srcs})
onnxruntime_add_include_to_target(onnxruntime_lora onnx flatbuffers::flatbuffers Boost::mp11 ${GSL_TARGET})
target_link_libraries(onnxruntime_lora onnxruntime_framework)
onnxruntime_add_include_to_target(onnxruntime_lora onnxruntime_framework onnxruntime_common onnx flatbuffers::flatbuffers Boost::mp11 ${GSL_TARGET})

if(onnxruntime_ENABLE_INSTRUMENT)
target_compile_definitions(onnxruntime_lora PUBLIC ONNXRUNTIME_ENABLE_INSTRUMENT)
Expand Down
5 changes: 5 additions & 0 deletions cmake/onnxruntime_providers_cuda.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -264,6 +264,11 @@
if("90" IN_LIST CMAKE_CUDA_ARCHITECTURES_ORIG)
target_compile_options(${target} PRIVATE $<$<COMPILE_LANGUAGE:CUDA>:-Xptxas=-w>)
target_compile_definitions(${target} PRIVATE COMPILE_HOPPER_TMA_GEMMS)
if (MSVC)
target_compile_options(${target} PRIVATE "$<$<COMPILE_LANGUAGE:CUDA>:SHELL:-Xcompiler /bigobj>")
target_compile_options(${target} PRIVATE "$<$<COMPILE_LANGUAGE:CUDA>:--diag-suppress=177>")
target_compile_options(${target} PRIVATE "$<$<COMPILE_LANGUAGE:CUDA>:SHELL:-Xcompiler /wd4172>")
endif()
endif()

if (onnxruntime_ENABLE_CUDA_PROFILING) # configure cupti for cuda profiling
Expand Down
1 change: 1 addition & 0 deletions cmake/onnxruntime_python.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -537,6 +537,7 @@ set(onnxruntime_mobile_util_srcs
${REPO_ROOT}/tools/python/util/pytorch_export_helpers.py
${REPO_ROOT}/tools/python/util/reduced_build_config_parser.py
${REPO_ROOT}/tools/python/util/update_onnx_opset.py
${REPO_ROOT}/tools/python/remove_initializer_from_input.py
)
file(GLOB onnxruntime_ort_format_model_srcs CONFIGURE_DEPENDS
${REPO_ROOT}/tools/python/util/ort_format_model/*.py
Expand Down
Loading
Loading