Skip to content

Commit

Permalink
fix typos (#446)
Browse files Browse the repository at this point in the history
  • Loading branch information
eric-haibin-lin committed Sep 28, 2020
1 parent 5412a33 commit 6f28ea3
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion deepspeed/runtime/pipe/p2p.py
Expand Up @@ -14,7 +14,7 @@ def init_process_groups(grid):
global _groups, _grid
_grid = grid

assert _grid.pipe_parallel_size > 1, "There is no model parallelism"
assert _grid.pipe_parallel_size > 1, "There is no pipeline parallelism"

_groups = [dist.new_group(ranks=group) for group in _grid.p2p_groups]

Expand Down
6 changes: 3 additions & 3 deletions deepspeed/runtime/zero/stage2.py
Expand Up @@ -260,7 +260,7 @@ def __init__(self,
for p, q in zip(self.fp16_groups[i], updated_params):
p.data = q.data

#divide the flat weights into near equal paritition equal to the data parallel degree
#divide the flat weights into near equal partition equal to the data parallel degree
#each process will compute on a different part of the partition
data_parallel_partitions = self.get_data_parallel_partitions(
self.fp16_groups_flat[i])
Expand Down Expand Up @@ -367,10 +367,10 @@ def __init__(self,
#stores the offset at which a parameter gradient needs to be inserted in a partition
self.grad_partition_insertion_offset = {}

#the offset in the gradient at which it must be inserted at the beginning of the paritition
#the offset in the gradient at which it must be inserted at the beginning of the partition
self.grad_start_offset = {}

#will store the averaged gradients required by this parititon
#will store the averaged gradients required by this partition
self.averaged_gradients = {}

# store index of first parameter in each partition
Expand Down

0 comments on commit 6f28ea3

Please sign in to comment.