-
Notifications
You must be signed in to change notification settings - Fork 25.6k
[Inductor] Add attention pattern for model DistilBert in transformers==4.44.2. #154091
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…==4.44.2. This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 Signed-off-by: xinan.lin <xinan.lin@intel.com> [ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/154091
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ✅ You can merge normally! (2 Unrelated Failures)As of commit 7a071e7 with merge base 6cb6da6 ( FLAKY - The following job failed but was likely due to flakiness present on trunk:
UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
…==4.44.2. This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 Signed-off-by: xinan.lin <xinan.linintel.com> ghstack-source-id: 1cdd0b3 Pull Request resolved: #154091
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
@pytorchbot revert -c ghfirst -m "I root caused this PR to some failures, I tried to resolve with #154923 but it looks like there are more failures with my fix" For a simple fix / follow-up to this all you have to do is re-run Sorry for the churn here this just made things difficult to land! |
@pytorchbot successfully started a revert job. Check the current status here. |
…sformers==4.44.2. (#154091)" This reverts commit 7b25ff7. Reverted #154091 on behalf of https://github.com/seemethere due to I root caused this PR to some failures, I tried to resolve with #154923 but it looks like there are more failures with my fix ([comment](#154091 (comment)))
@etaf your PR has been successfully reverted. |
…ransformers==4.44.2." This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov [ghstack-poisoned]
…ransformers==4.44.2." This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov [ghstack-poisoned]
…ransformers==4.44.2." This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov [ghstack-poisoned]
Starting merge as part of PR stack under #154110 |
Pull Request resolved: #154110 Approved by: https://github.com/eellison, https://github.com/jansel, https://github.com/EikanWang ghstack dependencies: #154091
…sformers==4.44.2. (pytorch#154091)" This reverts commit 7b25ff7. Reverted pytorch#154091 on behalf of https://github.com/seemethere due to I root caused this PR to some failures, I tried to resolve with pytorch#154923 but it looks like there are more failures with my fix ([comment](pytorch#154091 (comment)))
…==4.44.2. (pytorch#154091) This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 Pull Request resolved: pytorch#154091 Approved by: https://github.com/jansel, https://github.com/eellison
…54110) Pull Request resolved: pytorch#154110 Approved by: https://github.com/eellison, https://github.com/jansel, https://github.com/EikanWang ghstack dependencies: pytorch#154091
…==4.44.2. (pytorch#154091) This PR add a attention fusion pattern that match the attention of DistilDistilBert in transformers==4.44.2 at https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212 Pull Request resolved: pytorch#154091 Approved by: https://github.com/jansel, https://github.com/eellison
…54110) Pull Request resolved: pytorch#154110 Approved by: https://github.com/eellison, https://github.com/jansel, https://github.com/EikanWang ghstack dependencies: pytorch#154091
Stack from ghstack (oldest at bottom):
This PR add a attention fusion pattern that match the attention of
DistilDistilBert in transformers==4.44.2 at
https://github.com/huggingface/transformers/blob/953196a43dae6a3c474165fba7d215fcbc7b7730/src/transformers/models/distilbert/modeling_distilbert.py#L212
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @chenyang78 @kadeng @muchulee8 @amjames @chauhang @aakhundov