Skip to content

Commit c917c63

Browse files
naromero77amdpytorchmergebot
authored andcommitted
[ROCm][tunableop] UT tolerance increase for matmul_small_brute_force_tunableop at FP16 (pytorch#158788)
TunableOp will sometimes find a less precise solution due to the small input vectors used in this UT. Bumping op tolerance to eliminate flakiness. Pull Request resolved: pytorch#158788 Approved by: https://github.com/jeffdaily
1 parent 659bfbf commit c917c63

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

test/test_linalg.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4762,6 +4762,7 @@ def test_matmul_small_brute_force_3d_Nd(self, device, dtype):
47624762
@onlyCUDA
47634763
@skipCUDAIfNotRocm # Skipping due to SM89 OOM in CI, UT doesn't do much on NV anyways
47644764
@dtypes(*floating_types_and(torch.half))
4765+
@precisionOverride({torch.float16: 1e-1}) # TunableOp may occasionally find less precise solution
47654766
def test_matmul_small_brute_force_tunableop(self, device, dtype):
47664767
# disable tunableop buffer rotation for all tests everywhere, it can be slow
47674768
# We set the TunableOp numerical check environment variable here because it is

0 commit comments

Comments
 (0)