diff --git a/test/parallel/test_torch.py b/test/parallel/test_torch.py index b36dc38e3c..5212180cc6 100644 --- a/test/parallel/test_torch.py +++ b/test/parallel/test_torch.py @@ -631,6 +631,7 @@ def test_horovod_allreduce_duplicate_name_error(self): assert False, 'hvd.allreduce_async did not throw error' except (torch.FatalError, ValueError): pass + hvd.allreduce(torch.FloatTensor([0]), name="synch") def test_horovod_allreduce_grad(self): """Test the correctness of the allreduce gradient.""" @@ -1246,6 +1247,7 @@ def test_horovod_allgather_duplicate_name_error(self): assert False, 'hvd.allgather_async did not throw error' except (torch.FatalError, ValueError): pass + hvd.allreduce(torch.FloatTensor([0]), name="synch") def test_horovod_allgather_grad(self): """Test the correctness of the allgather gradient.""" @@ -1565,6 +1567,7 @@ def test_horovod_broadcast_duplicate_name_error(self): assert False, 'hvd.broadcast_async did not throw error' except (torch.FatalError, ValueError): pass + hvd.allreduce(torch.FloatTensor([0]), name="synch") def test_horovod_broadcast_grad(self): """Test the correctness of the broadcast gradient."""