Skip to content

Commit

Permalink
Bug fix to reset sequence parallelism (#6756) (#6770)
Browse files Browse the repository at this point in the history
* Bug fix to reset sequence parallelism



* Update seq par reset/restore



* Add nested loop



---------

Signed-off-by: Markel Sanz Ausin <markelsanz14@gmail.com>
Co-authored-by: Markel Sanz Ausin <markelsanz14@gmail.com>
  • Loading branch information
github-actions[bot] and markelsanz14 committed May 31, 2023
1 parent cd258bd commit 077a51b
Showing 1 changed file with 7 additions and 4 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -1274,12 +1274,13 @@ def _reset_sequence_parallelism_args(self):
self.last_sequence_parallel = self.cfg.sequence_parallel

# Reset config values. Needed for calling generate.
self.cfg.sequence_parallel = None
self.cfg.sequence_parallel = False

# Reset model parameters.

for module in self.get_gpt_module_list():
module.language_model.encoder.sequence_parallel = None
for mod in module.modules():
if hasattr(mod, "sequence_parallel"):
mod.sequence_parallel = self.last_sequence_parallel

def _restore_sequence_parallelism_args(self):
""" Restores the sequence parallelism parameters using the values saved by
Expand All @@ -1291,4 +1292,6 @@ def _restore_sequence_parallelism_args(self):

# Restore model parameters.
for module in self.get_gpt_module_list():
module.language_model.encoder.sequence_parallel = self.last_sequence_parallel
for mod in module.modules():
if hasattr(mod, "sequence_parallel"):
mod.sequence_parallel = self.last_sequence_parallel

0 comments on commit 077a51b

Please sign in to comment.