-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Distributed training all-reduce order #107
Comments
Marking as stale. No activity in 60 days. Remove stale label or comment or this will be closed in 7 days. |
Marking as stale. No activity in 60 days. |
I have encountered the same problem. Can anyone provide some opinions? |
@fwyc0573 are you seeing a hang? Can you describe the setting, perhaps provide an example command line, and also paste the last couple of lines in the logs? |
Marking as stale. No activity in 60 days. |
Hi,
I'm just wondering if there is a potential issue for all-reduce order when both data parallelism and tensor model parallelism are enabled during training. With torch DDP, both tensor model parallelism and data parallelism use all-reduce, and are launched on different streams. While the execution order is determined by hardware, will it cause hanging in some cases like:
GPU1: [MP] all-reduce -> [DP] all-reduce
GPU2: [DP] all-reduce -> [MP] all-reduce
From issues discussed here, I think it may be unsafe for undetermined order.
The text was updated successfully, but these errors were encountered: