Skip to content

Commit

Permalink
Merge branch 'distributed_indent_indentation' into 'main'
Browse files Browse the repository at this point in the history
Fix indentation for Torch distributed init_process_group

See merge request ADLR/megatron-lm!807
  • Loading branch information
jaredcasper committed Oct 7, 2023
2 parents 0c9c4f3 + faad056 commit f772743
Showing 1 changed file with 7 additions and 7 deletions.
14 changes: 7 additions & 7 deletions megatron/initialize.py
Original file line number Diff line number Diff line change
Expand Up @@ -192,13 +192,13 @@ def _initialize_distributed():
else:
args.local_rank = device
torch.cuda.set_device(device)
# Call the init process
torch.distributed.init_process_group(
backend=args.distributed_backend,
world_size=args.world_size,
rank=args.rank,
timeout=timedelta(minutes=args.distributed_timeout_minutes),
)
# Call the init process
torch.distributed.init_process_group(
backend=args.distributed_backend,
world_size=args.world_size,
rank=args.rank,
timeout=timedelta(minutes=args.distributed_timeout_minutes),
)

# Set the tensor model-parallel, pipeline model-parallel, and
# data-parallel communicators.
Expand Down

0 comments on commit f772743

Please sign in to comment.