From 72cd2aadc37bce4d2f7c8a7a05139d03c782aec1 Mon Sep 17 00:00:00 2001 From: Jintao Huang Date: Fri, 19 Jan 2024 14:31:42 +0800 Subject: [PATCH 1/2] fix freeze parameters --- swift/llm/utils/argument.py | 2 +- swift/utils/run_utils.py | 2 +- swift/utils/torch_utils.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/swift/llm/utils/argument.py b/swift/llm/utils/argument.py index b4c5117905..eba71b21a2 100644 --- a/swift/llm/utils/argument.py +++ b/swift/llm/utils/argument.py @@ -260,7 +260,7 @@ def __post_init__(self) -> None: and len( self.custom_val_dataset_path) == 0 and self.self_cognition_sample == 0): - raise ValueError(f'self.dataset: {self.dataset}') + raise ValueError(f'self.dataset: {self.dataset}, Please input the training dataset.') if self.save_steps is None: self.save_steps = self.eval_steps diff --git a/swift/utils/run_utils.py b/swift/utils/run_utils.py index e982df4fd6..ff7d62da27 100644 --- a/swift/utils/run_utils.py +++ b/swift/utils/run_utils.py @@ -15,7 +15,7 @@ def get_main( def x_main(argv: Union[List[str], _TArgsClass, NoneType] = None, **kwargs) -> _T: - if isinstance(argv, args_class): + if not isinstance(argv, (list, tuple, NoneType)): args, remaining_argv = argv, [] else: args, remaining_argv = parse_args(args_class, argv) diff --git a/swift/utils/torch_utils.py b/swift/utils/torch_utils.py index ee25793a85..b5fd8038b8 100644 --- a/swift/utils/torch_utils.py +++ b/swift/utils/torch_utils.py @@ -123,7 +123,7 @@ def show_layers(model: Module, max_lines: Optional[int] = 20) -> None: def freeze_model_parameters(model: Module, freeze_parameters: float) -> None: - n_parameters = np.array([p.numel() for p in model.parameters()]) + n_parameters = np.array([p.numel() for p in model.parameters()], dtype=np.int64) n_freeze_parameters = int(np.sum(n_parameters) * freeze_parameters) n_parameters_cs = np.cumsum(n_parameters) idx = bisect_right(n_parameters_cs, n_freeze_parameters) From ada1b8a1120d409f67e3863aef3baf7f5ee4eacd Mon Sep 17 00:00:00 2001 From: Jintao Huang Date: Fri, 19 Jan 2024 14:46:25 +0800 Subject: [PATCH 2/2] fix lint --- swift/llm/utils/argument.py | 4 +++- swift/utils/torch_utils.py | 3 ++- 2 files changed, 5 insertions(+), 2 deletions(-) diff --git a/swift/llm/utils/argument.py b/swift/llm/utils/argument.py index eba71b21a2..ac109aecb0 100644 --- a/swift/llm/utils/argument.py +++ b/swift/llm/utils/argument.py @@ -260,7 +260,9 @@ def __post_init__(self) -> None: and len( self.custom_val_dataset_path) == 0 and self.self_cognition_sample == 0): - raise ValueError(f'self.dataset: {self.dataset}, Please input the training dataset.') + raise ValueError( + f'self.dataset: {self.dataset}, Please input the training dataset.' + ) if self.save_steps is None: self.save_steps = self.eval_steps diff --git a/swift/utils/torch_utils.py b/swift/utils/torch_utils.py index b5fd8038b8..1ae8508fd8 100644 --- a/swift/utils/torch_utils.py +++ b/swift/utils/torch_utils.py @@ -123,7 +123,8 @@ def show_layers(model: Module, max_lines: Optional[int] = 20) -> None: def freeze_model_parameters(model: Module, freeze_parameters: float) -> None: - n_parameters = np.array([p.numel() for p in model.parameters()], dtype=np.int64) + n_parameters = np.array([p.numel() for p in model.parameters()], + dtype=np.int64) n_freeze_parameters = int(np.sum(n_parameters) * freeze_parameters) n_parameters_cs = np.cumsum(n_parameters) idx = bisect_right(n_parameters_cs, n_freeze_parameters)