NXP backend: Documentation for QAT#18228
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/18228
Note: Links to docs will display an error until the docs builds have been completed. ❌ 2 New Failures, 1 Pending, 2 Unrelated FailuresAs of commit a04917e with merge base faa5903 ( NEW FAILURES - The following jobs have failed:
FLAKY - The following jobs failed but were likely due to flakiness present on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
a7772d0 to
22d2206
Compare
|
Since we mention the |
22d2206 to
313bee6
Compare
|
@robert-kalmar @MartinPavella @roman-janik-nxp can some of you please take a look at this PR? |
There was a problem hiding this comment.
Pull request overview
Adds NXP (NeutronQuantizer) Quantization Aware Training (QAT) documentation and updates the NXP quantization helper to optionally run a training callback when is_qat=True.
Changes:
- Link NXP quantization docs from the quantization overview page.
- Add a new “Quantization Aware Training” section to NXP quantization documentation (workflow + limitations).
- Extend
calibrate_and_quantizeto support an optional QAT training callback and optionalcalibration_inputs.
Reviewed changes
Copilot reviewed 3 out of 3 changed files in this pull request and generated 7 comments.
| File | Description |
|---|---|
| docs/source/quantization-overview.md | Adds QAT/PTQ mention and links to NXP quantization docs. |
| docs/source/backends/nxp/nxp-quantization.md | Adds QAT documentation, examples, and known limitation/workaround notes. |
| backends/nxp/quantizer/utils.py | Adds optional QAT training via train_callback and makes calibration_inputs optional. |
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
I don't prefer that, as it introduce a new feature in the pipeline and the objective is, this PR to get cherry-picked to the release branch where new features are not strongly not preferred - #17016 (comment). Move the "training support" in separate PR to land only on main branch. The documentation is valid even without that change, if you remove the "Alternatively" part - https://github.com/pytorch/executorch/pull/18228/changes#diff-a510398a70db14eec60cc59d9f317799bc7aede898e9042c6e3ba00f4cac10cdR198 |
|
Okay, I will remove it. That's why I asked. |
313bee6 to
a04917e
Compare
|
LGTM. |
|
@pytorchbot cherry-pick --onto release/1.2 -c docs |
### Summary Adds documentation NeutronQuantizer support for QAT. (cherry picked from commit b1373e8)
Cherry picking #18228The cherry pick PR is at #18282 The following tracker issues are updated: Details for Dev Infra teamRaised by workflow job |
Summary
Adds documentation NeutronQuantizer support for QAT.
Adds train support for QAT mode in
calibrate_and_quantizefunction.Test plan
Mostly documentation related changes - no tests added.
The quantization related function is partially covered by already existing tests.
cc @robert-kalmar @JakeStevens @digantdesai