Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix TestOpInfoCUDA.test_unsupported_dtypes_addmm_cuda_bfloat16 on ampere #50440

Closed
wants to merge 1 commit into from

Conversation

zasdfgbnm
Copy link
Collaborator

The TestOpInfoCUDA.test_unsupported_dtypes_addmm_cuda_bfloat16 in test_ops.py is failing on ampere. This is because addmm is supported on Ampere, but the test is asserting that it is not supported.

@facebook-github-bot
Copy link
Contributor

facebook-github-bot commented Jan 12, 2021

💊 CI failures summary and remediations

As of commit 65ae5ec (more details on the Dr. CI page):


💚 💚 Looks good so far! There are no failures yet. 💚 💚


This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.

Please report bugs/suggestions to the (internal) Dr. CI Users group.

This comment has been revised 2 times.

@@ -785,7 +786,8 @@ def sample_inputs_fliplr_flipud(op_info, device, dtype, requires_grad):
OpInfo('addmm',
dtypes=floating_types(),
dtypesIfCPU=all_types_and_complex_and(torch.float16, torch.bfloat16),
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128),
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128,
*[torch.bfloat16] if tf32_is_not_fp32() else []),
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This will need a comment explaining the condition. Are we trying to test for Ampere or whether TF32 is enabled?

@mruberry mruberry added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label Jan 12, 2021
@@ -785,7 +786,8 @@ def sample_inputs_fliplr_flipud(op_info, device, dtype, requires_grad):
OpInfo('addmm',
dtypes=floating_types(),
dtypesIfCPU=all_types_and_complex_and(torch.float16, torch.bfloat16),
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128),
dtypesIfCUDA=floating_types_and(torch.float16, torch.complex64, torch.complex128,
*[torch.bfloat16] if tf32_is_not_fp32() else []),
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

That's a fun way of saying cuda11+ampere :-) Why doesn't cublas support (low perf) matrix multiplies on earlier architectures? (That's a rhetorical question).

@codecov
Copy link

codecov bot commented Jan 12, 2021

Codecov Report

Merging #50440 (65ae5ec) into master (5834438) will decrease coverage by 0.09%.
The diff coverage is 100.00%.

@@            Coverage Diff             @@
##           master   #50440      +/-   ##
==========================================
- Coverage   80.70%   80.61%   -0.10%     
==========================================
  Files        1905     1905              
  Lines      206813   206814       +1     
==========================================
- Hits       166916   166728     -188     
- Misses      39897    40086     +189     

Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ngimel has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@facebook-github-bot
Copy link
Contributor

@ngimel merged this pull request in 48318eb.

@zasdfgbnm zasdfgbnm deleted the matmul-testops branch January 13, 2021 20:04
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
cla signed Merged open source triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants