-
Notifications
You must be signed in to change notification settings - Fork 543
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Flow] Do not propagate reshape when it's blocking unpack+generic fusion #16930
Conversation
Abbreviated Benchmark Summary@ commit c2c1b4f9283fc60be7a91e7e2c09c461dcbdf9c5 (vs. base cc2ef92a232e4b6de9b845b6854d4d8667a6162b) Data-Tiling Comparison TableClick to show
No improved or regressed benchmarks 🏖️ Regressed Total Dispatch Sizes 🚩
[Top 3 out of 21 results showed] Improved Total Dispatch Sizes 🎉
[Top 3 out of 6 results showed] Regressed Total Artifact Sizes 🚩
Improved Stream IR Dispatch Count (# of cmd.dispatch ops) 🎉
[Top 3 out of 48 results showed] For more information: |
okay, so this does not only fix the regression, but also improves number of dispatches for other models. |
return isa<linalg::LinalgOp, tensor::UnPackOp>( | ||
reshapeOp.getSrc().getDefiningOp()); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we also add UnSetEncoding
op to the list?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice and simple change! Thanks!
It saves up to 20% number of dispatches in the benchmark suite.
Fixes #16835