Skip to content

Conversation

DrJessop
Copy link
Contributor

@DrJessop DrJessop commented Oct 9, 2025

Summary:
In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due.

This diff also fixes any ref implementations during the refactor.

Reviewed By: zonglinpeng

Differential Revision: D83873738

Copy link

pytorch-bot bot commented Oct 9, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/14974

Note: Links to docs will display an error until the docs builds have been completed.

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Oct 9, 2025
Copy link

meta-codesync bot commented Oct 9, 2025

@DrJessop has exported this pull request. If you are a Meta employee, you can view the originating Diff in D83873738.

Copy link

This PR needs a release notes: label

If your change should be included in the release notes (i.e. would users of this library care about this change?), please use a label starting with release notes:. This helps us keep track and include your important work in the next release notes.

To add a label, you can comment to pytorchbot, for example
@pytorchbot label "release notes: none"

For more information, see
https://github.com/pytorch/pytorch/wiki/PyTorch-AutoLabel-Bot#why-categorize-for-release-notes-and-how-does-it-work.

DrJessop pushed a commit to DrJessop/executorch that referenced this pull request Oct 10, 2025
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due.

This diff also fixes any ref implementations during the refactor.

Reviewed By: zonglinpeng

Differential Revision: D83873738
DrJessop pushed a commit to DrJessop/executorch that referenced this pull request Oct 10, 2025
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due.

This diff also fixes any ref implementations during the refactor.

Reviewed By: zonglinpeng

Differential Revision: D83873738
DrJessop pushed a commit to DrJessop/executorch that referenced this pull request Oct 10, 2025
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due.

This diff also fixes any ref implementations during the refactor.

Reviewed By: zonglinpeng

Differential Revision: D83873738
DrJessop pushed a commit to DrJessop/executorch that referenced this pull request Oct 10, 2025
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due.

This diff also fixes any ref implementations during the refactor.

Reviewed By: zonglinpeng

Differential Revision: D83873738
@DrJessop DrJessop force-pushed the export-D83873738 branch 2 times, most recently from f84f831 to 271538a Compare October 10, 2025 20:52
DrJessop pushed a commit to DrJessop/executorch that referenced this pull request Oct 10, 2025
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due. 

This diff directly replaces ops with the per-tensor variants and removes the pass which replaces singleton tensors with scalars.

Reviewed By: zonglinpeng

Differential Revision: D83873738
Andrew Grebenisan added 2 commits October 10, 2025 15:09
Summary:

Matmul was relying on linear infra which didn't support batched second argument. This adds support.

Differential Revision: D84279595
)

Summary:

In our previous flow, we would replace ops with default variants, have a special fusion pass which constructs singleton tensors for a variety of fused quantized ops, and then we would call a replace ops to turn them into per-tensor-variants.

I confirmed this was for legacy reasons, so a cleanup was much due. 

This diff directly replaces ops with the per-tensor variants and removes the pass which replaces singleton tensors with scalars.

Reviewed By: zonglinpeng

Differential Revision: D83873738
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant