Following AMD’s documented RyzenAI RAG workflow and using RyzenAI 1.7.0, the provided LLM examples do not run as provided.
https://github.com/amd/RyzenAI-SW/tree/main/LLM-examples/RAG-OGA
When running the RAG example exactly as documented (updating only the dataset path and model path as instructed), LLM initialization fails. The AMD-distributed LLM model configuration references the provider name RyzenAI, which results in the following error from onnxruntime-genai:
RuntimeError: Unknown provider name 'RyzenAI'
After correcting the provider name in the model configuration to VitisAI (as required by onnxruntime-genai), initialization fails again with:
RuntimeError: VitisAI execution provider is not supported in this build
This indicates that public onnxruntime-genai packages do not include NPU (VitisAI) support, even though:
RyzenAI installs NPU drivers successfully
AMD’s RAG examples and downloadable model configs imply that NPU-backed LLM inference should work
As a result, AMD’s downloadable files and documented directions do not work as-is. Users must either modify AMD-provided configuration files and then fall back to CPU execution, or obtain non-public builds to proceed.
