Skip to content

Commit

Permalink
Bug fix to restore act ckpt (NVIDIA#6753) (NVIDIA#6755)
Browse files Browse the repository at this point in the history
* Bug fix to restore act ckpt



* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Signed-off-by: Markel Sanz Ausin <markelsanz14@gmail.com>
Co-authored-by: Markel Sanz Ausin <markelsanz14@gmail.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Signed-off-by: hsiehjackson <c2hsieh@ucsd.edu>
  • Loading branch information
3 people authored and hsiehjackson committed Jun 2, 2023
1 parent 146371b commit 8f43ae3
Showing 1 changed file with 10 additions and 6 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -1248,16 +1248,20 @@ def _restore_activation_checkpointing_args(self):
_reset_activation_checkpointing_args.
"""
# Restore config values.
self.cfg.activations_checkpoint_granularity = self.last_checkpointing_granularity
self.cfg.activations_checkpoint_method = self.last_checkpointing_method
self.cfg.activations_checkpoint_num_layers = self.last_checkpointing_num_layers
self.cfg.activations_checkpoint_granularity = self.last_activations_checkpoint_granularity
self.cfg.activations_checkpoint_method = self.last_activations_checkpoint_method
self.cfg.activations_checkpoint_num_layers = self.last_activations_checkpoint_num_layers
self.cfg.activations_checkpoint_layers_per_pipeline = self.last_activations_checkpoint_layers_per_pipeline

# Restore model parameters.
for module in self.get_gpt_module_list():
module.language_model.encoder.activations_checkpoint_granularity = self.last_checkpointing_granularity
module.language_model.encoder.activations_checkpoint_method = self.last_checkpointing_method
module.language_model.encoder.activations_checkpoint_num_layers = self.last_checkpointing_num_layers
module.language_model.encoder.activations_checkpoint_granularity = (
self.last_activations_checkpoint_granularity
)
module.language_model.encoder.activations_checkpoint_method = self.last_activations_checkpoint_method
module.language_model.encoder.activations_checkpoint_num_layers = (
self.last_activations_checkpoint_num_layers
)
module.language_model.encoder.activations_checkpoint_layers_per_pipeline = (
self.last_activations_checkpoint_layers_per_pipeline
)
Expand Down

0 comments on commit 8f43ae3

Please sign in to comment.