-
Notifications
You must be signed in to change notification settings - Fork 24.6k
Move odd num_head in TransformerEncoder to slow_path #83483
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful links
✅ No Failures (0 Pending)As of commit 052ed6f (more details on the Dr. CI page): Expand to see more💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Please report bugs/suggestions to the (internal) Dr. CI Users group. |
This pull request was exported from Phabricator. Differential Revision: D38720086 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the fix for #83355!
Failure is because we did not add odd nhead check to creating a NestedTensor. So when nhead is odd, we are sending a NestedTensor to slow path, which cannot accept NestedTensor. Best fix is probably to check if the first or any encoder layer in TransformerEncoder has odd nhead. Kind of ugly... but what can you do. |
This pull request was exported from Phabricator. Differential Revision: D38720086 |
e234253
to
26c22a5
Compare
This pull request was exported from Phabricator. Differential Revision: D38720086 |
6eba5fe
to
fa96941
Compare
This pull request was exported from Phabricator. Differential Revision: D38720086 |
1 similar comment
This pull request was exported from Phabricator. Differential Revision: D38720086 |
fa96941
to
94cc7f1
Compare
This pull request was exported from Phabricator. Differential Revision: D38720086 |
94cc7f1
to
eff6f85
Compare
This pull request was exported from Phabricator. Differential Revision: D38720086 |
eff6f85
to
6cb96de
Compare
Summary: Pull Request resolved: pytorch#83483 odd nhead is not supported for masked softmax, therefore we just move it to use old slow_path Test Plan: CI Reviewed By: erichan1 Differential Revision: D38720086 fbshipit-source-id: 75abd450d6b6525b4c84470e5d189f9c498db753
This pull request was exported from Phabricator. Differential Revision: D38720086 |
6cb96de
to
052ed6f
Compare
@pytorchbot merge (Initiating merge automatically since Phabricator Diff has merged) |
@pytorchbot successfully started a merge job. Check the current status here. |
Hey @zrphercule. |
Summary: Pull Request resolved: #83483 odd nhead is not supported for masked softmax, therefore we just move it to use old slow_path Test Plan: CI Reviewed By: erichan1 Differential Revision: D38720086 fbshipit-source-id: 76d8745e8c5748ab0d2038b3ba0bc20fbbd05f9b
Summary: odd nhead is not supported for masked softmax, therefore we just move it to use old slow_path
Test Plan: CI
Differential Revision: D38720086