Skip to content

Conversation

@DannyYuyang-quic
Copy link
Contributor

Summary

Fix eval_llama_qnn: retrieve custom annotation from quantization recipe

Test plan

python -m executorch.examples.qualcomm.oss_scripts.llama.eval_llama_qnn --decoder_model qwen2_5-0_5b --quant_linear_only --max_seq_length 1024 --ptq 16a4w

@pytorch-bot
Copy link

pytorch-bot bot commented Nov 22, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/15953

Note: Links to docs will display an error until the docs builds have been completed.

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Nov 22, 2025
@DannyYuyang-quic
Copy link
Contributor Author

DannyYuyang-quic commented Nov 22, 2025

Hi @cccclai,

The previous PR #15807 moved all LLM quantization related configs into the quantization recipe.
As a result, custom_annotation was no longer accessible from the model config(LLMModelConfig), causing the unit test to fail.
https://github.com/pytorch/executorch/actions/runs/19558624238/job/56006215617#step:16:17963

This fix updates eval_llama_qnn.py to retrieve custom_annotation from the quantization recipe instead.

Please have a look, thanks!

@DannyYuyang-quic
Copy link
Contributor Author

@pytorchbot label "release notes: qualcomm"

@pytorch-bot pytorch-bot bot added the release notes: qualcomm Changes to the Qualcomm backend delegate label Nov 22, 2025
@cccclai
Copy link
Contributor

cccclai commented Nov 22, 2025

lint is failing, can you fix it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. release notes: qualcomm Changes to the Qualcomm backend delegate

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants