diff --git a/auto_round/compressors/base.py b/auto_round/compressors/base.py index 8f398e7a1..7ab11d0ff 100644 --- a/auto_round/compressors/base.py +++ b/auto_round/compressors/base.py @@ -2699,7 +2699,7 @@ def _quantize_block( ) logger.info(dump_info) if self.low_gpu_mem_usage: - clear_memory(self.device_list) # clear cached memory during training + clear_memory(device_list=self.device_list) # clear cached memory during training if len(unquantized_layer_names) != 0: logger.info(f"{unquantized_layer_names} have not been quantized") with torch.no_grad():