Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 3 additions & 4 deletions test/quantization/test_qat.py
Original file line number Diff line number Diff line change
Expand Up @@ -96,6 +96,7 @@
get_groupwise_affine_qparams,
groupwise_affine_quantize_tensor,
)
from torchao.testing.utils import skip_if_xpu
from torchao.utils import (
_is_fbgemm_gpu_genai_available,
get_current_accelerator_device,
Expand Down Expand Up @@ -695,10 +696,7 @@ def test_qat_4w_quantizer_gradients(self):
self._test_qat_quantized_gradients(quantizer)

@unittest.skipIf(_DEVICE is None, "skipping when GPU is not available")
@unittest.skipIf(
_DEVICE is torch.device("xpu"),
"skipped due to https://github.com/intel/torch-xpu-ops/issues/1770",
)
@skip_if_xpu("skipped due to https://github.com/intel/torch-xpu-ops/issues/1770")
def test_qat_4w_quantizer(self):
from torchao.quantization.GPTQ import Int4WeightOnlyQuantizer
from torchao.quantization.qat import Int4WeightOnlyQATQuantizer
Expand Down Expand Up @@ -2015,6 +2013,7 @@ def test_quantize_api_int8_intx(self, weight_dtype, weight_granularity, dtype):
)

@unittest.skipIf(_DEVICE is None, "skipping when GPU is not available")
@skip_if_xpu("XPU enablement in progress")
@parametrize(
"weight_dtype, granularity, dtype, module_type",
[
Expand Down
Loading