From f94dd06d58fa912d44fa2ce7cba971d2949dd2f3 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Wed, 13 Aug 2025 08:25:27 -0700 Subject: [PATCH 1/3] Bump keras in /models_v2/tensorflow/bert_large/inference/cpu (#2808) Bumps [keras](https://github.com/keras-team/keras) from 3.10.0 to 3.11.0. - [Release notes](https://github.com/keras-team/keras/releases) - [Commits](https://github.com/keras-team/keras/compare/v3.10.0...v3.11.0) --- updated-dependencies: - dependency-name: keras dependency-version: 3.11.0 dependency-type: indirect ... Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Srikanth Ramakrishna --- .../bert_large/inference/cpu/poetry.lock | 26 +++++++++---------- 1 file changed, 13 insertions(+), 13 deletions(-) diff --git a/models_v2/tensorflow/bert_large/inference/cpu/poetry.lock b/models_v2/tensorflow/bert_large/inference/cpu/poetry.lock index ce304e87a..822e0299c 100644 --- a/models_v2/tensorflow/bert_large/inference/cpu/poetry.lock +++ b/models_v2/tensorflow/bert_large/inference/cpu/poetry.lock @@ -1,4 +1,4 @@ -# This file is automatically @generated by Poetry 2.0.0 and should not be changed by hand. +# This file is automatically @generated by Poetry 2.1.1 and should not be changed by hand. [[package]] name = "absl-py" @@ -302,14 +302,14 @@ all = ["flake8 (>=7.1.1)", "mypy (>=1.11.2)", "pytest (>=8.3.2)", "ruff (>=0.6.2 [[package]] name = "keras" -version = "3.11.0" +version = "3.11.2" description = "Multi-backend Keras" optional = false python-versions = ">=3.10" groups = ["main"] files = [ - {file = "keras-3.11.0-py3-none-any.whl", hash = "sha256:28babe8dc69370a2b7c3bc385be6c1f22d702b00b3c47cae62559b6f80d6073a"}, - {file = "keras-3.11.0.tar.gz", hash = "sha256:f5dfeaf4fcaea180e032f7c1e373f1868961e2940dcfcaaf9a5b711baf41bd60"}, + {file = "keras-3.11.2-py3-none-any.whl", hash = "sha256:539354b1870dce22e063118c99c766c3244030285b5100b4a6f8840145436bf0"}, + {file = "keras-3.11.2.tar.gz", hash = "sha256:b78a4af616cbe119e88fa973d2b0443b70c7f74dd3ee888e5026f0b7e78a2801"}, ] [package.dependencies] @@ -503,9 +503,9 @@ files = [ [package.dependencies] numpy = [ {version = ">=2.1.0", markers = "python_version >= \"3.13\""}, - {version = ">=1.26.0", markers = "python_version >= \"3.12\" and python_version < \"3.13\""}, - {version = ">=1.23.3", markers = "python_version >= \"3.11\" and python_version < \"3.12\""}, - {version = ">=1.21.2", markers = "python_version >= \"3.10\" and python_version < \"3.11\""}, + {version = ">=1.26.0", markers = "python_version == \"3.12\""}, + {version = ">=1.23.3", markers = "python_version == \"3.11\""}, + {version = ">=1.21.2", markers = "python_version == \"3.10\""}, ] [package.extras] @@ -797,13 +797,13 @@ files = [ ] [package.extras] -check = ["pytest-checkdocs (>=2.4)", "pytest-ruff (>=0.2.1)", "ruff (>=0.8.0)"] -core = ["importlib_metadata (>=6)", "jaraco.functools (>=4)", "jaraco.text (>=3.7)", "more_itertools", "more_itertools (>=8.8)", "packaging (>=24.2)", "platformdirs (>=4.2.2)", "tomli (>=2.0.1)", "wheel (>=0.43.0)"] +check = ["pytest-checkdocs (>=2.4)", "pytest-ruff (>=0.2.1) ; sys_platform != \"cygwin\"", "ruff (>=0.8.0) ; sys_platform != \"cygwin\""] +core = ["importlib_metadata (>=6) ; python_version < \"3.10\"", "jaraco.functools (>=4)", "jaraco.text (>=3.7)", "more_itertools", "more_itertools (>=8.8)", "packaging (>=24.2)", "platformdirs (>=4.2.2)", "tomli (>=2.0.1) ; python_version < \"3.11\"", "wheel (>=0.43.0)"] cover = ["pytest-cov"] doc = ["furo", "jaraco.packaging (>=9.3)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "pyproject-hooks (!=1.1)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (>=1,<2)", "sphinx-reredirects", "sphinxcontrib-towncrier", "towncrier (<24.7)"] enabler = ["pytest-enabler (>=2.2)"] -test = ["build[virtualenv] (>=1.0.3)", "filelock (>=3.4.0)", "ini2toml[lite] (>=0.14)", "jaraco.develop (>=7.21)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.7.2)", "jaraco.test (>=5.5)", "packaging (>=24.2)", "pip (>=19.1)", "pyproject-hooks (!=1.1)", "pytest (>=6,!=8.1.*)", "pytest-home (>=0.5)", "pytest-perf", "pytest-subprocess", "pytest-timeout", "pytest-xdist (>=3)", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel (>=0.44.0)"] -type = ["importlib_metadata (>=7.0.2)", "jaraco.develop (>=7.21)", "mypy (==1.14.*)", "pytest-mypy"] +test = ["build[virtualenv] (>=1.0.3)", "filelock (>=3.4.0)", "ini2toml[lite] (>=0.14)", "jaraco.develop (>=7.21) ; python_version >= \"3.9\" and sys_platform != \"cygwin\"", "jaraco.envs (>=2.2)", "jaraco.path (>=3.7.2)", "jaraco.test (>=5.5)", "packaging (>=24.2)", "pip (>=19.1)", "pyproject-hooks (!=1.1)", "pytest (>=6,!=8.1.*)", "pytest-home (>=0.5)", "pytest-perf ; sys_platform != \"cygwin\"", "pytest-subprocess", "pytest-timeout", "pytest-xdist (>=3)", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel (>=0.44.0)"] +type = ["importlib_metadata (>=7.0.2) ; python_version < \"3.10\"", "jaraco.develop (>=7.21) ; sys_platform != \"cygwin\"", "mypy (==1.14.*)", "pytest-mypy"] [[package]] name = "six" @@ -913,7 +913,7 @@ description = "TensorFlow IO" optional = false python-versions = "<3.13,>=3.7" groups = ["main"] -markers = "python_version < \"3.12\"" +markers = "python_version == \"3.11\" or python_version == \"3.10\"" files = [ {file = "tensorflow_io_gcs_filesystem-0.37.1-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:249c12b830165841411ba71e08215d0e94277a49c551e6dd5d72aab54fe5491b"}, {file = "tensorflow_io_gcs_filesystem-0.37.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:257aab23470a0796978efc9c2bcf8b0bc80f22e6298612a4c0a50d3f4e88060c"}, @@ -980,7 +980,7 @@ files = [ ] [package.extras] -brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)"] +brotli = ["brotli (>=1.0.9) ; platform_python_implementation == \"CPython\"", "brotlicffi (>=0.8.0) ; platform_python_implementation != \"CPython\""] h2 = ["h2 (>=4,<5)"] socks = ["pysocks (>=1.5.6,!=1.5.7,<2.0)"] zstd = ["zstandard (>=0.18.0)"] From 9bbcf7ea6ffd5f814150b141fdd7a8564fac726b Mon Sep 17 00:00:00 2001 From: Xia Weiwen Date: Wed, 3 Sep 2025 23:59:47 +0800 Subject: [PATCH 2/3] Turn on use_small_dequant_buffer flag for llm imference on CPU with Inductor (#2810) --- .../llama/inference/cpu/inductor/run_llm_inductor_greedy.py | 1 + 1 file changed, 1 insertion(+) diff --git a/models_v2/pytorch/llama/inference/cpu/inductor/run_llm_inductor_greedy.py b/models_v2/pytorch/llama/inference/cpu/inductor/run_llm_inductor_greedy.py index b50f878d8..9161e1f8c 100644 --- a/models_v2/pytorch/llama/inference/cpu/inductor/run_llm_inductor_greedy.py +++ b/models_v2/pytorch/llama/inference/cpu/inductor/run_llm_inductor_greedy.py @@ -119,6 +119,7 @@ def trace_handler(prof): inductor_config.cpp_wrapper = True inductor_config.max_autotune = True inductor_config.max_autotune_gemm_backends = "CPP,ATEN" +inductor_config.cpp.use_small_dequant_buffer = True torch._dynamo.config.allow_unspec_int_on_nn_module = True if args.dtype in ["fp32","bf16", "fp16"]: From 40b047d476994af5549f17d6aa8869c21b230abb Mon Sep 17 00:00:00 2001 From: Srikanth Ramakrishna Date: Thu, 11 Sep 2025 17:17:10 -0700 Subject: [PATCH 3/3] replace extra-index-url with index-irl (#2819) --- docs/general/pytorch/BareMetalSetup.md | 2 +- models_v2/pytorch/bert_large/inference/gpu/README.md | 2 +- models_v2/pytorch/bert_large/training/gpu/README.md | 2 +- models_v2/pytorch/distilbert/inference/gpu/README.md | 2 +- models_v2/pytorch/resnet50v1_5/inference/gpu/README.md | 2 +- models_v2/pytorch/resnet50v1_5/training/gpu/README.md | 2 +- models_v2/pytorch/rnnt/inference/gpu/README.md | 2 +- models_v2/pytorch/rnnt/training/gpu/README.md | 2 +- models_v2/pytorch/ssd-mobilenetv1/inference/gpu/README.md | 2 +- .../pytorch/stable_diffusion/inference/gpu/README.md | 2 +- models_v2/pytorch/torchrec_dlrm/inference/gpu/README.md | 2 +- notebooks/AI_Reference_Models.ipynb | 8 ++++---- tests/cicd/PyTorch/setup.sh | 2 +- tests/cicd/pytorch/pyt_cpu_setup.sh | 2 +- tests/cicd/pytorch/pyt_xpu_setup.sh | 2 +- 15 files changed, 18 insertions(+), 18 deletions(-) diff --git a/docs/general/pytorch/BareMetalSetup.md b/docs/general/pytorch/BareMetalSetup.md index 018ca89b1..ddd86583f 100644 --- a/docs/general/pytorch/BareMetalSetup.md +++ b/docs/general/pytorch/BareMetalSetup.md @@ -8,7 +8,7 @@ Prepare the environment, you may create a Python virtual enviromment `virtualenv python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu # Install oneccl: - python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ + python -m pip install oneccl_bind_pt --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ ## Install generic dependencies Make sure the following components are installed in your environment : diff --git a/models_v2/pytorch/bert_large/inference/gpu/README.md b/models_v2/pytorch/bert_large/inference/gpu/README.md index 3c931887f..e149030db 100644 --- a/models_v2/pytorch/bert_large/inference/gpu/README.md +++ b/models_v2/pytorch/bert_large/inference/gpu/README.md @@ -67,7 +67,7 @@ BERT_WEIGHT=/squad_large_finetuned_checkpoint ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/bert_large/training/gpu/README.md b/models_v2/pytorch/bert_large/training/gpu/README.md index 1a9f64f5c..017bc9222 100644 --- a/models_v2/pytorch/bert_large/training/gpu/README.md +++ b/models_v2/pytorch/bert_large/training/gpu/README.md @@ -53,7 +53,7 @@ We are using `hdf5/hdf5_4320_shards_varlength` as the our dataset. ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Setup required environment paramaters diff --git a/models_v2/pytorch/distilbert/inference/gpu/README.md b/models_v2/pytorch/distilbert/inference/gpu/README.md index e8c720505..c9b27fca7 100644 --- a/models_v2/pytorch/distilbert/inference/gpu/README.md +++ b/models_v2/pytorch/distilbert/inference/gpu/README.md @@ -48,7 +48,7 @@ the dataset should be like below ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/resnet50v1_5/inference/gpu/README.md b/models_v2/pytorch/resnet50v1_5/inference/gpu/README.md index 0e173d963..09d2265d3 100644 --- a/models_v2/pytorch/resnet50v1_5/inference/gpu/README.md +++ b/models_v2/pytorch/resnet50v1_5/inference/gpu/README.md @@ -41,7 +41,7 @@ ImageNet is recommended, the download link is https://image-net.org/challenges/L ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/resnet50v1_5/training/gpu/README.md b/models_v2/pytorch/resnet50v1_5/training/gpu/README.md index 712efd5d2..06fa9cbc7 100644 --- a/models_v2/pytorch/resnet50v1_5/training/gpu/README.md +++ b/models_v2/pytorch/resnet50v1_5/training/gpu/README.md @@ -40,7 +40,7 @@ ImageNet is recommended, the download link is https://image-net.org/challenges/L ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/rnnt/inference/gpu/README.md b/models_v2/pytorch/rnnt/inference/gpu/README.md index c853fe395..7594b1622 100644 --- a/models_v2/pytorch/rnnt/inference/gpu/README.md +++ b/models_v2/pytorch/rnnt/inference/gpu/README.md @@ -80,7 +80,7 @@ Specify the folder which contains the `rnnt_ckpt.pt` model as parameter `WEIGHT_ ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/rnnt/training/gpu/README.md b/models_v2/pytorch/rnnt/training/gpu/README.md index dfc1c5e1e..665205b1f 100644 --- a/models_v2/pytorch/rnnt/training/gpu/README.md +++ b/models_v2/pytorch/rnnt/training/gpu/README.md @@ -79,7 +79,7 @@ please specify the folder path which contains datasets/LibriSpeech as parameter ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/ssd-mobilenetv1/inference/gpu/README.md b/models_v2/pytorch/ssd-mobilenetv1/inference/gpu/README.md index 0473ebecc..1814a8155 100644 --- a/models_v2/pytorch/ssd-mobilenetv1/inference/gpu/README.md +++ b/models_v2/pytorch/ssd-mobilenetv1/inference/gpu/README.md @@ -55,7 +55,7 @@ Download from [this URL](https://drive.google.com/drive/folders/1pKn-RifvJGWiOx0 ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/stable_diffusion/inference/gpu/README.md b/models_v2/pytorch/stable_diffusion/inference/gpu/README.md index bbcbd311d..c82967a5f 100644 --- a/models_v2/pytorch/stable_diffusion/inference/gpu/README.md +++ b/models_v2/pytorch/stable_diffusion/inference/gpu/README.md @@ -39,7 +39,7 @@ The scripts will download the dataset automatically. It uses nateraw/parti-promp ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/models_v2/pytorch/torchrec_dlrm/inference/gpu/README.md b/models_v2/pytorch/torchrec_dlrm/inference/gpu/README.md index 382cb9c05..833f0aee2 100644 --- a/models_v2/pytorch/torchrec_dlrm/inference/gpu/README.md +++ b/models_v2/pytorch/torchrec_dlrm/inference/gpu/README.md @@ -57,7 +57,7 @@ and the folder will be used as the parameter WEIGHT_DIR later ``` 5. Install the latest GPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation): ``` - python -m pip install torch== torchvision== intel-extension-for-pytorch== --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch== torchvision== intel-extension-for-pytorch== --index-url https://pytorch-extension.intel.com/release-whl-aitools/ ``` 6. Set environment variables for Intel® oneAPI Base Toolkit: Default installation location `{ONEAPI_ROOT}` is `/opt/intel/oneapi` for root account, `${HOME}/intel/oneapi` for other accounts diff --git a/notebooks/AI_Reference_Models.ipynb b/notebooks/AI_Reference_Models.ipynb index 9fc17c984..8da57ce41 100644 --- a/notebooks/AI_Reference_Models.ipynb +++ b/notebooks/AI_Reference_Models.ipynb @@ -168,7 +168,7 @@ " if selected_hardware == 'CPU':\n", " !python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu\n", " !python -m pip install intel-extension-for-pytorch\n", - " !python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/\n", + " !python -m pip install oneccl_bind_pt --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/\n", "\n", " print(\"PyTorch CPU packages installed.\")\n", "\n", @@ -190,7 +190,7 @@ " if (install_oneapi):\n", " print(\"Intel® oneAPI Base Toolkit is installed\")\n", " print(\"Installing PyTorch for GPU...\")\n", - " !pip install torch==2.1.0a0 torchvision==0.16.a0 torchaudio==2.1.0.a0 intel-extension-for-pytorch==2.1.10+xpu --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/\n", + " !pip install torch==2.1.0a0 torchvision==0.16.a0 torchaudio==2.1.0.a0 intel-extension-for-pytorch==2.1.10+xpu --index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/\n", " print(\"IPEX GPU packages installed.\")\n", "\n", " if gpu_series == 'Flex':\n", @@ -803,7 +803,7 @@ " if selected_hardware == 'CPU':\n", " !python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu\n", " !python -m pip install intel-extension-for-pytorch\n", - " !python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/\n", + " !python -m pip install oneccl_bind_pt --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/\n", "\n", " print(\"PyTorch CPU packages installed.\")\n", "\n", @@ -825,7 +825,7 @@ " if (install_oneapi):\n", " print(\"Intel® oneAPI Base Toolkit is installed\")\n", " print(\"Installing PyTorch for GPU...\")\n", - " !pip install torch==2.1.0a0 torchvision==0.16.a0 torchaudio==2.1.0.a0 intel-extension-for-pytorch==2.1.10+xpu --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/\n", + " !pip install torch==2.1.0a0 torchvision==0.16.a0 torchaudio==2.1.0.a0 intel-extension-for-pytorch==2.1.10+xpu --index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/\n", " print(\"IPEX GPU packages installed.\")\n", "\n", " if gpu_series == 'Flex':\n", diff --git a/tests/cicd/PyTorch/setup.sh b/tests/cicd/PyTorch/setup.sh index c95a25065..c4d4ce917 100755 --- a/tests/cicd/PyTorch/setup.sh +++ b/tests/cicd/PyTorch/setup.sh @@ -48,7 +48,7 @@ else export NO_PROXY="" python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu python -m pip install intel-extension-for-pytorch - python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ + python -m pip install oneccl_bind_pt --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ fi # Check the operating system type diff --git a/tests/cicd/pytorch/pyt_cpu_setup.sh b/tests/cicd/pytorch/pyt_cpu_setup.sh index fcd798454..f585f2bb5 100755 --- a/tests/cicd/pytorch/pyt_cpu_setup.sh +++ b/tests/cicd/pytorch/pyt_cpu_setup.sh @@ -27,7 +27,7 @@ else export NO_PROXY=“” python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu python -m pip install intel-extension-for-pytorch - python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ + python -m pip install oneccl_bind_pt --index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ fi # Check the operating system type diff --git a/tests/cicd/pytorch/pyt_xpu_setup.sh b/tests/cicd/pytorch/pyt_xpu_setup.sh index 986db4107..522e6cb1f 100755 --- a/tests/cicd/pytorch/pyt_xpu_setup.sh +++ b/tests/cicd/pytorch/pyt_xpu_setup.sh @@ -26,5 +26,5 @@ else echo "Installing pytorch" export no_proxy=“” export NO_PROXY=“” - python -m pip install torch==${FRAMEWORK_VERSION} torchvision==${TORCHVISION_VERSION} intel-extension-for-pytorch==${FRAMEWORK_EXTENSION_VERSION} --extra-index-url https://pytorch-extension.intel.com/release-whl-aitools/ + python -m pip install torch==${FRAMEWORK_VERSION} torchvision==${TORCHVISION_VERSION} intel-extension-for-pytorch==${FRAMEWORK_EXTENSION_VERSION} --index-url https://pytorch-extension.intel.com/release-whl-aitools/ fi