Skip to content
This repository has been archived by the owner on Nov 22, 2022. It is now read-only.

support using RoundRobin ProcessGroup in Distributed training #1213

Conversation

chenyangyu1988
Copy link
Contributor

Summary:
support using RoundRobin ProcessGroup in Distributed training
RoundRobin ProcessGroup will use multi-stream for gradient sync.

Usually it could give us 15%+ speed up with no gradient accumulation.

Reviewed By: hudeven

Differential Revision: D19138726

Summary:
support using RoundRobin ProcessGroup in Distributed training
RoundRobin ProcessGroup will use multi-stream for gradient sync.

Usually it could give us 15%+ speed up with no gradient accumulation.

Reviewed By: hudeven

Differential Revision: D19138726

fbshipit-source-id: bc52df1ebeb9bd69a5239507dab866533a56f6b5
@facebook-github-bot facebook-github-bot added CLA Signed Do not delete this pull request or issue due to inactivity. fb-exported labels Dec 20, 2019
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D19138726

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 151be72.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
CLA Signed Do not delete this pull request or issue due to inactivity. fb-exported Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants