We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Please answer the following questions for yourself before submitting an issue.
Intel® Core™ Ultra processors now has released , how can llama.cpp use that npu to fast up
Intel® Core™ Ultra processors deliver three dedicated engines (CPU, GPU, and NPU) to help unlock the power of AI https://www.intel.com/content/www/us/en/products/docs/processors/core-ultra/core-ultra-series-1-product-brief.html
OpenVINO™, WindowsML, DirectML, ONNX RT
The text was updated successfully, but these errors were encountered:
Intel open source its NPU implementation library. Looks like it's using OpenVINO
https://github.com/intel/intel-npu-acceleration-library https://intel.github.io/intel-npu-acceleration-library/cpp_reference.html
Sorry, something went wrong.
This issue was closed because it has been inactive for 14 days since being marked as stale.
Hi, any plans for this?
No branches or pull requests
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Feature Description
Intel® Core™ Ultra processors now has released , how can llama.cpp use that npu to fast up
Motivation
Intel® Core™ Ultra processors deliver three dedicated engines (CPU, GPU, and NPU) to help unlock the power of AI
https://www.intel.com/content/www/us/en/products/docs/processors/core-ultra/core-ultra-series-1-product-brief.html
Possible Implementation
OpenVINO™, WindowsML, DirectML, ONNX RT
The text was updated successfully, but these errors were encountered: