@@ -2576,7 +2576,7 @@ def test_fused_attention_different_dk_dv(self, device):
2576
2576
2577
2577
2578
2578
@skipIfRocm # No cuDNN Attention
2579
- @unittest .skipIf (not PLATFORM_SUPPORTS_CUDNN_ATTENTION , "cuDNN Attention is not supported on this system " )
2579
+ @unittest .skipIf (True , "broken as of cuDNN 9.10 " )
2580
2580
def test_cudnn_attention_fail_d128 (self , device ):
2581
2581
# Test that cuDNN attention dispatching correctly bails out on d > 128
2582
2582
b , h = 1 , 2
@@ -2591,7 +2591,6 @@ def test_cudnn_attention_fail_d128(self, device):
2591
2591
ISSM90 = device_cap == (9 , 0 )
2592
2592
ISSM100 = device_cap == (10 , 0 )
2593
2593
with sdpa_kernel (backends = [SDPBackend .CUDNN_ATTENTION ]):
2594
- # SM90/100 support d <= 256 as of cuDNN 9.5.1+
2595
2594
if (ISSM90 or ISSM100 ) and torch .backends .cudnn .version () >= 90501 :
2596
2595
torch .nn .functional .scaled_dot_product_attention (q , k , v )
2597
2596
else :
@@ -3030,7 +3029,9 @@ def test_fused_sdp_choice(self, device, type: str):
3030
3029
device_capability = None
3031
3030
if "cuda" in str (device ):
3032
3031
device_capability = torch .cuda .get_device_capability ()
3033
- prefer_cudnn = device_capability and (device_capability == (9 , 0 ) or device_capability == (10 , 0 ))
3032
+ prefer_cudnn = False
3033
+ # TODO(eqy): uncomment the following condition
3034
+ # device_capability and (device_capability == (9, 0) or device_capability == (10, 0))
3034
3035
3035
3036
# TODO we are currently disabling this by default, lets assert that this returns
3036
3037
# FlashAttention, we need to change when we make remove opt-in for cudnn
0 commit comments