From deb59e1a8634668a789a6c3a1753f06ec6720c73 Mon Sep 17 00:00:00 2001 From: Jithun Nair Date: Thu, 14 Feb 2019 17:20:45 +0000 Subject: [PATCH 1/2] Re-disabling test_cudnn_multiple_threads_same_device for ROCm as it is flaky on CI --- test/test_nn.py | 1 + 1 file changed, 1 insertion(+) diff --git a/test/test_nn.py b/test/test_nn.py index 5adf5b5a249d5..c3761bcfb53cd 100644 --- a/test/test_nn.py +++ b/test/test_nn.py @@ -3858,6 +3858,7 @@ def test_Conv2d_inconsistent_types_on_GPU_with_cudnn(self): @unittest.skipIf(not TEST_CUDA, 'CUDA not available') @unittest.skipIf(not TEST_CUDNN, 'CUDNN not available') + @skipIfRocm def test_cudnn_multiple_threads_same_device(self): # This function is intended to test the lazy creation and reuse of per-thread # cudnn handles on each device in aten/src/ATen/cudnn/Handles.cpp. From bcabdd92759ea7d6639c0323134f64a7cab58c68 Mon Sep 17 00:00:00 2001 From: Jithun Nair Date: Thu, 14 Feb 2019 19:00:48 +0000 Subject: [PATCH 2/2] Disable test_alexnet since it fails on CI as well --- test/test_jit.py | 1 + 1 file changed, 1 insertion(+) diff --git a/test/test_jit.py b/test/test_jit.py index 8847b51924426..0e677b4d60ea4 100644 --- a/test/test_jit.py +++ b/test/test_jit.py @@ -1341,6 +1341,7 @@ def fn(a, b): self.assertExpectedGraph(ge.graph) @skipIfNoTorchVision + @skipIfRocm def test_alexnet(self): x = torch.ones(1, 3, 224, 224) trace, _ = torch.jit.get_trace_graph(torchvision.models.AlexNet(), x)