Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
49 commits
Select commit Hold shift + click to select a range
2f8d79e
Remove build-nuget from dml-vs-2022.yml (#24372)
snnn Apr 10, 2025
f44f1a9
Make test `CApiTest.RequestLoadCancellation` deterministic (#24348)
yuslepukhin Apr 10, 2025
d58ff6b
[web] allow NPM tests to run nodejs binding for webgpu (#24370)
fs-eire Apr 11, 2025
7a03764
MlasTranspose multi-threads support. (#24261)
msy-kato Apr 11, 2025
2d5316f
[webgpu] Use workgroup memory to reduce register pressure (#24286)
qjia7 Apr 11, 2025
f7028a3
[QNN EP] Add support for int64 shape input of Expand Op (#24389)
quic-tirupath Apr 11, 2025
c70425a
[webgpu] fix 2 bugs in Conv/ConvTranspose (#24388)
fs-eire Apr 11, 2025
d003b6e
Migrate OpenVino Pipeline to Github Actions (#24297)
snnn Apr 11, 2025
aada488
[Native WebGPU EP] Add InstranceNormalization (#24369)
satyajandhyala Apr 11, 2025
d7c688e
Bump vite from 6.2.5 to 6.2.6 in /js/web/test/e2e/exports/testcases/v…
dependabot[bot] Apr 11, 2025
699b4db
Update protobuf-java to 3.25.5 (#24333)
jchen351 Apr 12, 2025
90c263f
Add API to compile a model (#24207)
adrianlizarraga Apr 12, 2025
9a993c3
[CPU] Add 8bit support to matmulnbits quantizer (#24384)
fajin-corp Apr 14, 2025
bbd2739
[webgpu] move comments out from WGSL in FlashAttention impl (#24400)
fs-eire Apr 14, 2025
7417f43
Replace gsl::narrow with narrow in xnnpack code (#24392)
cdliang11 Apr 14, 2025
bb5a879
[WebGPU EP] Fixes bugs in slice operator implementation (#24415)
prathikr Apr 14, 2025
d205bb7
Support mixed precision in quantization for RTN (#24401)
jiafatom Apr 14, 2025
1f14dac
Add Resize cubic mode without antialias (scales = [1, ≥1, ≥1, 1]) (#2…
yihonglyu Apr 15, 2025
8de1639
[webgpu] Enable DP4A MatMul generation path for Qualcomm (#24408)
qjia7 Apr 15, 2025
ac5e434
workaround linux CI pipeline: pin triton to v3.2.0 (#24423)
fs-eire Apr 15, 2025
fd22509
Fix typo in option text s/buildings/bindings (#24412)
clementperon Apr 15, 2025
c47d694
[Native WebGPU EP] Increase error tolerance limit for f16 (#24420)
satyajandhyala Apr 15, 2025
c619218
Fix doc gen issue (#24424)
HectorSVC Apr 15, 2025
c18e06d
[MacOS] Add MLProgram Gather op for CoreML EP (#24387)
carzh Apr 15, 2025
ff607b4
[Native WebGPU] Support shared memory version of ReduceOps (#24399)
satyajandhyala Apr 15, 2025
fc7595f
[web] fix 'npm run pull:wasm' for main branch (#24429)
fs-eire Apr 15, 2025
c27975f
Fix the Python API docs update pipeline (#24434)
HectorSVC Apr 15, 2025
4b24ddf
[node.js] fix handling null value for externalData (#24428)
fs-eire Apr 15, 2025
d60891d
[webgpu] Fix batch-norm for ort-web-tests (#24404)
jchen10 Apr 15, 2025
98f075c
Enable Inference Results Saving in onnx-test-runner (#24210)
quic-hungjuiw Apr 15, 2025
cf99ca9
[nodejs] allow installing DLLs from Nuget feed (#24418)
fs-eire Apr 15, 2025
39d60b7
Clean up Compile API (#24436)
adrianlizarraga Apr 16, 2025
4d03aef
Support canonical EP names in SessionOptionsAppendExecutionProvider (…
adrianlizarraga Apr 16, 2025
4b9e26d
Add static quantization runner (#24114)
quic-hungjuiw Apr 16, 2025
ec02406
[nodejs] support Node.js binding in multi env (#24366)
fs-eire Apr 16, 2025
b4b5a79
[QNN EP] Enable QnnGpu backend in QNN EP. (#24435)
johnpaultaken Apr 16, 2025
201012e
Update QNN version to 2.33.2 (#24440)
HectorSVC Apr 16, 2025
171e297
ONNXRuntime OpenVINO - Release 1.22 (#24394)
preetha-intel Apr 16, 2025
9ab6b87
Add GQA fusion for CUDA EP (#24335)
nenad1002 Apr 16, 2025
42c2a62
[nodejs] upgrade N-API version to 6 (#24443)
fs-eire Apr 16, 2025
01c8e83
Fix compile issue in Azure EP unit test (#24446)
adrianlizarraga Apr 16, 2025
1c2225e
[WebNN EP] Automatically use ml-tensor for outputs (#24282)
egalli Apr 16, 2025
c5b82a5
[nodejs] add missing header files for linux build (#24448)
fs-eire Apr 16, 2025
f267b7e
Fix MatmulTransposeFusion when input A and B are the same (#24373)
fs-eire Apr 17, 2025
99f2b80
Fix cuda memory access violation in GQA FlashAttention (#24447)
RyanUnderhill Apr 17, 2025
ef832b9
[QNN EP] Reverting a recent logging change for QNN GPU only, (#24444)
johnpaultaken Apr 17, 2025
9db4b10
[nodejs] update Node.js binding document for 1.22 release (#24452)
fs-eire Apr 17, 2025
3252320
[Native WebGPU] Handle corner cases in naive kernel. (#24438)
satyajandhyala Apr 17, 2025
254bda0
Merge branch 'master' into sync_msft_17_4_25
jatinwadhwa921 Apr 17, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
45 changes: 45 additions & 0 deletions .github/workflows/linux_openvino_ci.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
name: Linux OpenVINO CI

on:
push:
branches: [ main, 'rel-*' ]
pull_request:
branches: [ main, 'rel-*' ]
workflow_dispatch:

concurrency:
group: ${{ github.workflow }}-${{ github.ref }}
cancel-in-progress: true

permissions:
contents: read
packages: write # Needed if the reusable workflow pushes images
attestations: write # Optional: for artifact attestations if enabled
id-token: write # Optional: may be needed for OIDC authentication (e.g., ACR)

jobs:
build_test_openvino:
name: Build and Test OpenVINO EP (AlamLinux8, Py3.12)
# Use the reusable workflow as the other Linux CI pipelines
uses: ./.github/workflows/reusable_linux_build.yml
with:
pool_name: "onnxruntime-github-Ubuntu2204-AMD-CPU"
build_config: Release
# Architecture: OpenVino only supports Intel X64
architecture: x64
dockerfile_path: tools/ci_build/github/linux/docker/inference/x86_64/python/openvino/Dockerfile
docker_image_repo: onnxruntimeopenvino

execution_providers: 'openvino'

extra_build_flags: '--use_openvino CPU --enable_generic_interface --build_shared_lib'

# Python Path Prefix: Set the correct Python 3.12 path inside the manylinux container
python_path_prefix: 'PATH=/opt/python/cp312-cp312/bin:$PATH'

run_tests: true
upload_build_output: false

# Secrets: Pass the necessary GitHub token
secrets:
GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}
1 change: 1 addition & 0 deletions .github/workflows/publish-python-apidocs.yml
Original file line number Diff line number Diff line change
Expand Up @@ -40,6 +40,7 @@ jobs:
- name: Generate Python docs with Sphinx
run: |
cd tools/doc
chmod +x *
./builddoc.sh /usr/bin ../.. ../../build
- name: Log source commit
run: git rev-parse --short HEAD > build/docs/html/source-version.txt
Expand Down
2 changes: 1 addition & 1 deletion cgmanifests/cgmanifest.json
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
"maven": {
"GroupId": "com.google.protobuf",
"ArtifactId": "protobuf-java",
"Version": "3.21.7"
"Version": "3.25.5"
},
"DevelopmentDependency": true
}
Expand Down
2 changes: 1 addition & 1 deletion cmake/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ option(onnxruntime_GENERATE_TEST_REPORTS "Enable test report generation" OFF)
option(onnxruntime_ENABLE_STATIC_ANALYSIS "Enable static analysis" OFF)
option(onnxruntime_USE_CUSTOM_STATIC_ANALYSIS_RULES "Use a custom SDL Rule. It is mainly for our CI build" OFF)
option(onnxruntime_REDIRECT_STATIC_ANALYSIS_OUTPUTS_TO_FILE "Use a custom SDL Rule. It is mainly for our CI build" OFF)
option(onnxruntime_ENABLE_PYTHON "Enable python buildings" OFF)
option(onnxruntime_ENABLE_PYTHON "Enable python bindings" OFF)
# Enable it may cause LNK1169 error
option(onnxruntime_ENABLE_MEMLEAK_CHECKER "Experimental: Enable memory leak checker in Windows debug build" OFF)
option(onnxruntime_USE_CUDA "Build with CUDA support" OFF)
Expand Down
4 changes: 2 additions & 2 deletions docs/python/ReadMeOV.rst
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ Requirements
^^^^^^^^^^^^

- Ubuntu 18.04, 20.04 or Windows 10 - 64 bit
- Python 3.11, 3.12 and 3.13 for Windows and Linux
- Python 3.10, 3.11, 3.12 and 3.13 for Windows and Linux

This package supports:
- Intel® CPUs
Expand All @@ -29,7 +29,7 @@ This package supports:
Please install OpenVINO™ PyPi Package separately for Windows.
For installation instructions on Windows please refer to `OpenVINO™ Execution Provider for ONNX Runtime for Windows <https://github.com/intel/onnxruntime/releases/>`_.

**OpenVINO™ Execution Provider for ONNX Runtime** Linux Wheels comes with pre-built libraries of OpenVINO™ version 2025.0.0 eliminating the need to install OpenVINO™ separately.
**OpenVINO™ Execution Provider for ONNX Runtime** Linux Wheels comes with pre-built libraries of OpenVINO™ version 2025.1.0 eliminating the need to install OpenVINO™ separately.

For more details on build and installation please refer to `Build <https://onnxruntime.ai/docs/build/eps.html#openvino>`_.

Expand Down
19 changes: 19 additions & 0 deletions include/onnxruntime/core/providers/providers.h
Original file line number Diff line number Diff line change
Expand Up @@ -2,12 +2,31 @@
// Licensed under the MIT License.

#pragma once
#include <memory>

struct OrtSessionOptions;
struct OrtLogger;

namespace onnxruntime {
class IExecutionProvider;

struct IExecutionProviderFactory {
virtual ~IExecutionProviderFactory() = default;
virtual std::unique_ptr<IExecutionProvider> CreateProvider() = 0;

/// <summary>
/// Creates an IExecutionProvider instance. Enables initialization of an EP instance using session-level options
/// such as session configs (string key/value pairs), graph optimization level, etc.
///
/// The default implementation ignores the arguments and calls the above CreateProvider() function,
/// which does not take in any arguments.
///
/// This version of CreateProvider() is used by InferenceSession when registering EPs.
/// </summary>
/// <param name="session_options">Options for the session in which the IExecutionProvider is used</param>
/// <param name="session_logger">Session logger that should be used by the IExecutionProvider.</param>
/// <returns>An IExecutionProvider</returns>
virtual std::unique_ptr<IExecutionProvider> CreateProvider(const OrtSessionOptions& session_options,
const OrtLogger& session_logger);
};
} // namespace onnxruntime
Loading
Loading