Skip to content

Conversation

navsud
Copy link
Contributor

@navsud navsud commented Sep 24, 2025

Summary: To save GPU memory bfloat16 dtype is commonly used for training of LLMs. Currently, the quantizer ignores quantizing the nodes if they are not float32. This change enables quantization of bf16 nodes as well.

Differential Revision: D82866443

@navsud navsud requested a review from cccclai as a code owner September 24, 2025 22:00
Copy link

pytorch-bot bot commented Sep 24, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/14558

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

❌ 4 New Failures, 2 Cancelled Jobs

As of commit 9a886d2 with merge base c98079a (image):

NEW FAILURES - The following jobs have failed:

CANCELLED JOBS - The following jobs were cancelled. Please retry:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Sep 24, 2025
@facebook-github-bot
Copy link
Contributor

@navsud has exported this pull request. If you are a Meta employee, you can view the originating diff in D82866443.

@navsud navsud added the release notes: none Do not include this in the release notes label Sep 24, 2025
navsud added a commit to navsud/executorch that referenced this pull request Sep 24, 2025
Summary:

To save GPU memory `bfloat16` dtype is commonly used for training of LLMs. Currently, the quantizer ignores quantizing the nodes if they are not float32. This change enables quantization of bf16 nodes as well.

Differential Revision: D82866443
Summary:

To save GPU memory `bfloat16` dtype is commonly used for training of LLMs. Currently, the quantizer ignores quantizing the nodes if they are not float32. This change enables quantization of bf16 nodes as well.

Reviewed By: billmguo

Differential Revision: D82866443
@facebook-github-bot
Copy link
Contributor

@navsud has exported this pull request. If you are a Meta employee, you can view the originating diff in D82866443.

@facebook-github-bot facebook-github-bot merged commit 2283294 into pytorch:main Sep 26, 2025
125 of 132 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported release notes: none Do not include this in the release notes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants