From aae4b1352fef1ff7d0d9b843c35aca39a524d92f Mon Sep 17 00:00:00 2001 From: akashveramd Date: Mon, 7 Jul 2025 14:54:34 -0700 Subject: [PATCH 1/2] Cherry-picked commit with merge conflict --- test/distributed/test_c10d_gloo.py | 11 +++++++++++ 1 file changed, 11 insertions(+) diff --git a/test/distributed/test_c10d_gloo.py b/test/distributed/test_c10d_gloo.py index aec59c276698e..44b88f76b340d 100644 --- a/test/distributed/test_c10d_gloo.py +++ b/test/distributed/test_c10d_gloo.py @@ -53,7 +53,11 @@ retry_on_connect_failures, run_tests, skip_but_pass_in_sandcastle, +<<<<<<< HEAD skipIfRocmArch, +======= + skipIfRocm, +>>>>>>> 2269e37502 (Skipped *_stress_cuda UTs in test_c10d_gloo in release/2.5 branch. (#2317)) TestCase, ) @@ -387,6 +391,7 @@ def test_broadcast_stress(self): inputs = [torch.tensor([i * self.world_size + self.rank]) for i in range(1000)] self._test_broadcast_stress(inputs) + @skipIfRocm @skip_if_lt_x_gpu(2) @requires_gloo() def test_broadcast_stress_cuda(self): @@ -492,6 +497,7 @@ def test_allreduce_stress(self): inputs = [torch.tensor([i + self.rank]) for i in range(1000)] self._test_allreduce_stress(inputs) + @skipIfRocm @skip_if_lt_x_gpu(2) @requires_gloo() def test_allreduce_stress_cuda(self): @@ -924,6 +930,8 @@ def test_scatter_stress(self): @skip_but_pass_in_sandcastle( "Test is flaky, see https://github.com/pytorch/pytorch/issues/15963" ) + + @skipIfRocm @skip_if_lt_x_gpu(2) @requires_gloo() def test_scatter_stress_cuda(self): @@ -1098,6 +1106,7 @@ def test_gather_stress(self): inputs = [torch.tensor([i + self.rank]) for i in range(1000)] self._test_gather_stress(inputs, lambda t: t.clone()) + @skipIfRocm @skip_if_lt_x_gpu(2) @skipIfRocmArch(MI300_ARCH) @requires_gloo() @@ -1234,6 +1243,7 @@ def test_allgather_stress(self): inputs = [torch.tensor([i + self.rank]) for i in range(1000)] self._test_allgather_stress(inputs, lambda t: t.clone()) + @skipIfRocm @skip_if_lt_x_gpu(2) @requires_gloo() def test_allgather_stress_cuda(self): @@ -1420,6 +1430,7 @@ def test_reduce_stress(self): inputs = [torch.tensor([i + self.rank]) for i in range(1000)] self._test_reduce_stress(inputs) + @skipIfRocm @skip_if_lt_x_gpu(2) @requires_gloo() def test_reduce_stress_cuda(self): From 6b403cddf378a3e114cbece072b8d22aa6abd2f9 Mon Sep 17 00:00:00 2001 From: Prachi Gupta Date: Mon, 28 Jul 2025 17:29:24 +0000 Subject: [PATCH 2/2] Skipped *_stress_cuda UTs in test_c10d_gloo --- test/distributed/test_c10d_gloo.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/test/distributed/test_c10d_gloo.py b/test/distributed/test_c10d_gloo.py index 44b88f76b340d..90a4b4e1ab4a9 100644 --- a/test/distributed/test_c10d_gloo.py +++ b/test/distributed/test_c10d_gloo.py @@ -53,11 +53,8 @@ retry_on_connect_failures, run_tests, skip_but_pass_in_sandcastle, -<<<<<<< HEAD skipIfRocmArch, -======= skipIfRocm, ->>>>>>> 2269e37502 (Skipped *_stress_cuda UTs in test_c10d_gloo in release/2.5 branch. (#2317)) TestCase, )