-
Notifications
You must be signed in to change notification settings - Fork 22.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix TestOpInfoCUDA.test_unsupported_dtypes_addmm_cuda_bfloat16 on ampere #50440
Conversation
💊 CI failures summary and remediationsAs of commit 65ae5ec (more details on the Dr. CI page): 💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.Please report bugs/suggestions to the (internal) Dr. CI Users group. This comment has been revised 2 times. |
@@ -785,7 +786,8 @@ def sample_inputs_fliplr_flipud(op_info, device, dtype, requires_grad): | |||
OpInfo('addmm', | |||
dtypes=floating_types(), | |||
dtypesIfCPU=all_types_and_complex_and(torch.float16, torch.bfloat16), | |||
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128), | |||
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128, | |||
*[torch.bfloat16] if tf32_is_not_fp32() else []), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This will need a comment explaining the condition. Are we trying to test for Ampere or whether TF32 is enabled?
@@ -785,7 +786,8 @@ def sample_inputs_fliplr_flipud(op_info, device, dtype, requires_grad): | |||
OpInfo('addmm', | |||
dtypes=floating_types(), | |||
dtypesIfCPU=all_types_and_complex_and(torch.float16, torch.bfloat16), | |||
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128), | |||
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128, | |||
*[torch.bfloat16] if tf32_is_not_fp32() else []), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That's a fun way of saying cuda11+ampere :-) Why doesn't cublas support (low perf) matrix multiplies on earlier architectures? (That's a rhetorical question).
Codecov Report
@@ Coverage Diff @@
## master #50440 +/- ##
==========================================
- Coverage 80.70% 80.61% -0.10%
==========================================
Files 1905 1905
Lines 206813 206814 +1
==========================================
- Hits 166916 166728 -188
- Misses 39897 40086 +189 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@ngimel has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
The
TestOpInfoCUDA.test_unsupported_dtypes_addmm_cuda_bfloat16
intest_ops.py
is failing on ampere. This is because addmm is supported on Ampere, but the test is asserting that it is not supported.