From 2e50ad354c091f4a35d6605ddd78c46d747a3d10 Mon Sep 17 00:00:00 2001 From: Dmitry Nikolaev <139769634+dnikolaev-amd@users.noreply.github.com> Date: Sat, 19 Jul 2025 07:26:15 +0200 Subject: [PATCH 1/2] Cherry-picked commit with merge conflict --- test/test_nn.py | 8 ++++++++ 1 file changed, 8 insertions(+) diff --git a/test/test_nn.py b/test/test_nn.py index d5c245c5887d2..24a02e900b929 100644 --- a/test/test_nn.py +++ b/test/test_nn.py @@ -5174,6 +5174,7 @@ def test_batchnorm_buffer_update_when_stats_are_not_tracked(self): name_fn=lambda f, b, m, t: f"{f}_vs_{b}{'_mixed' if m else ''}_{dtype_name(t)}" ) def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype): +<<<<<<< HEAD if torch.version.cuda: if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16", "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16"): @@ -5182,6 +5183,10 @@ def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype): if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16": self.skipTest("Batchnorm 3D NHWC train failed on CUDA") +======= + if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16": + self.skipTest("3D float16 NCHW train failed on CUDA and ROCm due to Native batchnorm accuracy issue SWDEV-541024") +>>>>>>> 4eaa5bf23b ([rocm7.0_internal_testing] skip 3D NCHW FP16 batchnorm test due to Native accuracy issue (#2370)) if torch.version.hip: if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16", "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16") \ @@ -5197,10 +5202,13 @@ def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype): # https://github.com/pytorch/pytorch/issues/156513 self.skipTest("bfloat16 NCHW train failed due to native tolerance issue") +<<<<<<< HEAD if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16" \ and _get_torch_rocm_version() < (7, 0): self.skipTest("3D float16 NCHW train failed on ROCm<7.0") +======= +>>>>>>> 4eaa5bf23b ([rocm7.0_internal_testing] skip 3D NCHW FP16 batchnorm test due to Native accuracy issue (#2370)) if dims == 3 and memory_format in ("NHWC", "NCHW"): memory_format = memory_format + "3D" From cc4843880d0970415586e33c078cbd759e592c32 Mon Sep 17 00:00:00 2001 From: Dmitry Nikolaev <139769634+dnikolaev-amd@users.noreply.github.com> Date: Wed, 5 Nov 2025 21:33:39 +0100 Subject: [PATCH 2/2] Update test_nn.py resolve conflict --- test/test_nn.py | 29 +++++++++++------------------ 1 file changed, 11 insertions(+), 18 deletions(-) diff --git a/test/test_nn.py b/test/test_nn.py index 24a02e900b929..0c84d6ffe129e 100644 --- a/test/test_nn.py +++ b/test/test_nn.py @@ -5174,25 +5174,22 @@ def test_batchnorm_buffer_update_when_stats_are_not_tracked(self): name_fn=lambda f, b, m, t: f"{f}_vs_{b}{'_mixed' if m else ''}_{dtype_name(t)}" ) def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype): -<<<<<<< HEAD if torch.version.cuda: if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16", - "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16"): - self.skipTest("bfloat16 NHWC train failed on CUDA due to native tolerance issue " - "https://github.com/pytorch/pytorch/issues/156513") - if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16": - self.skipTest("Batchnorm 3D NHWC train failed on CUDA") + "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16", + "test_batchnorm_2D_train_NHWC_vs_NCHW_mixed_bfloat16", + "test_batchnorm_3D_train_NHWC_vs_NCHW_mixed_bfloat16", + "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16"): + self.skipTest("Failed on CUDA") -======= - if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16": - self.skipTest("3D float16 NCHW train failed on CUDA and ROCm due to Native batchnorm accuracy issue SWDEV-541024") ->>>>>>> 4eaa5bf23b ([rocm7.0_internal_testing] skip 3D NCHW FP16 batchnorm test due to Native accuracy issue (#2370)) if torch.version.hip: if self._testMethodName in ("test_batchnorm_2D_train_NCHW_vs_cpu_mixed_bfloat16", - "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16") \ + "test_batchnorm_3D_train_NCHW_vs_cpu_mixed_bfloat16", + "test_batchnorm_2D_train_NHWC_vs_NCHW_mixed_bfloat16", + "test_batchnorm_3D_train_NHWC_vs_NCHW_mixed_bfloat16") \ and _get_torch_rocm_version() < (6, 4): # NCHW bfloat16 path uses native kernels for rocm<=6.3 - # train failed on rocm<=6.3 due to native tolerance issue + # train failed on rocm<=6.3 due to native accuracy issue # https://github.com/pytorch/pytorch/issues/156513 self.skipTest("bfloat16 NHWC train failed on ROCm <= 6.3") @@ -5202,13 +5199,9 @@ def test_batchnorm(self, dims, mode, memory_format, ref_backend, mixed, dtype): # https://github.com/pytorch/pytorch/issues/156513 self.skipTest("bfloat16 NCHW train failed due to native tolerance issue") -<<<<<<< HEAD - if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16" \ - and _get_torch_rocm_version() < (7, 0): - self.skipTest("3D float16 NCHW train failed on ROCm<7.0") + if self._testMethodName == "test_batchnorm_3D_train_NCHW_vs_native_mixed_float16": + self.skipTest("3D float16 NCHW train failed on ROCm") -======= ->>>>>>> 4eaa5bf23b ([rocm7.0_internal_testing] skip 3D NCHW FP16 batchnorm test due to Native accuracy issue (#2370)) if dims == 3 and memory_format in ("NHWC", "NCHW"): memory_format = memory_format + "3D"