Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[c10d] Ensure collectives are called with the same dtype for all tensor params. #84664

Closed
wants to merge 5 commits into from

Conversation

kumpera
Copy link
Contributor

@kumpera kumpera commented Sep 7, 2022

While passing tensors with different dtypes don't crash, they don't produce sensible results.

We see data tearing instead of casting.

It's not clear we want to support transparent casting so, for now, we fail when such input is presented.

Fixes #84525

Fixes #ISSUE_NUMBER

@pytorch-bot
Copy link

pytorch-bot bot commented Sep 7, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/84664

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 7d31ac2:
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the release notes: distributed (c10d) release notes category label Sep 7, 2022
@facebook-github-bot facebook-github-bot added cla signed oncall: distributed Add this issue/PR to distributed oncall triage queue labels Sep 7, 2022
@numpee
Copy link

numpee commented Sep 8, 2022

Thanks for addressing the issue. This would be a great QoL update

Rodrigo Kumpera added 3 commits September 12, 2022 18:59
…or params.

While passing tensors with different dtypes don't crash, they don't produce sensible results.

We see data tearing instead of casting.

It's not clear we want to support transparent casting so, for now, we fail when such input
is presented.

Fixes pytorch#84525
Copy link
Member

@rohan-varma rohan-varma left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM overall, but I guess it is technically BC breaking?

torch/distributed/distributed_c10d.py Show resolved Hide resolved
@kumpera
Copy link
Contributor Author

kumpera commented Sep 15, 2022

@pytorchmergebot merge

@kumpera kumpera added topic: bc breaking topic category topic: improvements topic category labels Sep 15, 2022
@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a merge job. Check the current status here and land check progress here.
The merge job was triggered with the land checks (-l) flag. If you did not specify this flag yourself, you are likely enrolled in the land checks rollout. This means that your change will be merged once all checks on your PR and the land checks have passed (ETA 4 Hours). If you need to coordinate lands between different changes and cannot risk a land race, please add the ciflow/trunk label to your PR and wait for signal to complete, and then land your changes in proper order. Having trunk, pull, and Lint pre-run on a PR will bypass land checks and the ETA should be immediate. If this is not the intended behavior, feel free to use some of the other merge options in the wiki.
Please reach out to the PyTorch DevX Team with feedback or questions!

pytorchmergebot pushed a commit that referenced this pull request Sep 15, 2022
…or params. (#84664)

While passing tensors with different dtypes don't crash, they don't produce sensible results.

We see data tearing instead of casting.

It's not clear we want to support transparent casting so, for now, we fail when such input is presented.

Fixes #84525

Fixes #ISSUE_NUMBER

Pull Request resolved: #84664
Approved by: https://github.com/rohan-varma
mehtanirav pushed a commit that referenced this pull request Oct 4, 2022
…or params. (#84664)

While passing tensors with different dtypes don't crash, they don't produce sensible results.

We see data tearing instead of casting.

It's not clear we want to support transparent casting so, for now, we fail when such input is presented.

Fixes #84525

Fixes #ISSUE_NUMBER

Pull Request resolved: #84664
Approved by: https://github.com/rohan-varma
alvgaona pushed a commit to alvgaona/pytorch that referenced this pull request Oct 11, 2022
…or params. (pytorch#84664)

While passing tensors with different dtypes don't crash, they don't produce sensible results.

We see data tearing instead of casting.

It's not clear we want to support transparent casting so, for now, we fail when such input is presented.

Fixes pytorch#84525

Fixes #ISSUE_NUMBER

Pull Request resolved: pytorch#84664
Approved by: https://github.com/rohan-varma
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
cla signed Merged oncall: distributed Add this issue/PR to distributed oncall triage queue release notes: distributed (c10d) release notes category topic: bc breaking topic category topic: improvements topic category
Projects
None yet
Development

Successfully merging this pull request may close these issues.

torch.distributed.all_gather on wrong type of tensor list should raise a TypeError
5 participants