From c388620f6c17d310fb1201a79ccdcfd15a6efd74 Mon Sep 17 00:00:00 2001 From: Jesse Cai Date: Fri, 31 Oct 2025 15:31:04 -0400 Subject: [PATCH 1/2] Set model parameters to requires_grad=False Disable gradient computation for model parameters to fix test failing after https://github.com/pytorch/pytorch/pull/166367 --- test/sparsity/test_marlin.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/test/sparsity/test_marlin.py b/test/sparsity/test_marlin.py index e602210ee5..3cd4e5f2ab 100644 --- a/test/sparsity/test_marlin.py +++ b/test/sparsity/test_marlin.py @@ -39,6 +39,9 @@ def setUp(self): .half() .cuda() ) + for param in self.model.parameters(): + param.requires_grad=False + @pytest.mark.skipif(not torch.cuda.is_available(), reason="Need CUDA available") @skip_if_rocm("ROCm enablement in progress") From 456f1f214a8ca015736594036637d970d1559dac Mon Sep 17 00:00:00 2001 From: Jesse Cai Date: Fri, 31 Oct 2025 15:33:29 -0400 Subject: [PATCH 2/2] Fix formatting of requires_grad assignment --- test/sparsity/test_marlin.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/test/sparsity/test_marlin.py b/test/sparsity/test_marlin.py index 3cd4e5f2ab..dd07c31172 100644 --- a/test/sparsity/test_marlin.py +++ b/test/sparsity/test_marlin.py @@ -40,8 +40,7 @@ def setUp(self): .cuda() ) for param in self.model.parameters(): - param.requires_grad=False - + param.requires_grad = False @pytest.mark.skipif(not torch.cuda.is_available(), reason="Need CUDA available") @skip_if_rocm("ROCm enablement in progress")