From 3fd6dd147681f1f521dadf0255b1eeb70582bf57 Mon Sep 17 00:00:00 2001 From: Chen Lai Date: Tue, 23 Apr 2024 18:08:38 -0700 Subject: [PATCH] fix qnn install link (#3260) Summary: Pull Request resolved: https://github.com/pytorch/executorch/pull/3260 As title, the link was wrong... Reviewed By: kirklandsign Differential Revision: D56498322 fbshipit-source-id: 42708b5f7a634f1c01e05af4c897d0c6da54d724 (cherry picked from commit e9d7868abd2e5cd9aa5b6e91c5dc22ed757cc0bd) --- examples/models/llama2/export_llama_lib.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/examples/models/llama2/export_llama_lib.py b/examples/models/llama2/export_llama_lib.py index eef0bcee965..26f684af09c 100644 --- a/examples/models/llama2/export_llama_lib.py +++ b/examples/models/llama2/export_llama_lib.py @@ -662,7 +662,8 @@ def _export_llama(modelname, args) -> str: # noqa: C901 quantizers = [] except ImportError: raise ImportError( - "Please install the Qualcomm backend follwing https://pytorch.org/executorch/main/build-run-qualcomm.html" + "Please install the Qualcomm backend follwing " + "https://pytorch.org/executorch/main/build-run-qualcomm-ai-engine-direct-backend.html" ) backend, quant_config = args.pt2e_quantize.split("_")