Skip to content

Conversation

fduwjj
Copy link
Contributor

@fduwjj fduwjj commented Oct 13, 2023

Stack from ghstack (oldest at bottom):

As part of TP UX improvements, we want to keep our API simple (not easy) so that users get the flexibility to do what they want and avoid a too generic API which tries to solve everything and get things too complicated. We are updating the doc accordingly.

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 13, 2023

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/111176

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 9f3b9d7 with merge base 35750bf (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

fduwjj added a commit that referenced this pull request Oct 13, 2023
@fduwjj fduwjj requested a review from wanchaol October 13, 2023 01:35
@fduwjj fduwjj changed the title Remove documentations for Pairwise parallel and sequence parallel [TP] Remove documentations for Pairwise parallel, sequence parallel and other prepare input/output functions Oct 13, 2023
@fduwjj fduwjj added ciflow/trunk Trigger trunk jobs on your pull request module: dtensor distributed tensor tag release notes: distributed (dtensor) release notes category labels Oct 13, 2023
Copy link
Collaborator

@wanchaol wanchaol left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

looks good

… parallel and other prepare input/output functions"


As part of TP UX improvements, we want to keep our API simple (not easy) so that users get the flexibility to do what they want and avoid a too generic API which tries to solve everything and get things too complicated. We are updating the doc accordingly.


[ghstack-poisoned]
fduwjj added a commit that referenced this pull request Oct 13, 2023
… parallel and other prepare input/output functions"


As part of TP UX improvements, we want to keep our API simple (not easy) so that users get the flexibility to do what they want and avoid a too generic API which tries to solve everything and get things too complicated. We are updating the doc accordingly.


[ghstack-poisoned]
… parallel and other prepare input/output functions"


As part of TP UX improvements, we want to keep our API simple (not easy) so that users get the flexibility to do what they want and avoid a too generic API which tries to solve everything and get things too complicated. We are updating the doc accordingly.


[ghstack-poisoned]
fduwjj added a commit that referenced this pull request Oct 14, 2023
@wanchaol
Copy link
Collaborator

https://docs-preview.pytorch.org/pytorch/pytorch/111176/distributed.tensor.parallel.html looks like prepare input/outout docs still have issues

@fduwjj
Copy link
Contributor Author

fduwjj commented Oct 15, 2023

@pytorchbot merge

@fduwjj fduwjj changed the title [TP] Remove documentations for Pairwise parallel, sequence parallel and other prepare input/output functions [TP] Add deprecation warnings in the documentations for Pairwise parallel, sequence parallel and other prepare input/output functions Oct 15, 2023
@fduwjj
Copy link
Contributor Author

fduwjj commented Oct 15, 2023

@wanchaol Let me add warnings for them in #111177.

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@wanchaol
Copy link
Collaborator

@fduwjj This is not related to deprecation warning, see here https://docs-preview.pytorch.org/pytorch/pytorch/111176/distributed.tensor.parallel.html#torch.distributed.tensor.parallel.style.PrepareModuleOutput

The examples in the doc are not properly formatted

pytorchmergebot pushed a commit that referenced this pull request Oct 15, 2023
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.

Pull Request resolved: #111177
Approved by: https://github.com/wanchaol
ghstack dependencies: #111160, #111166, #111176
pytorchmergebot pushed a commit that referenced this pull request Oct 16, 2023
yeounoh pushed a commit to yeounoh/pytorch that referenced this pull request Oct 16, 2023
…llel, sequence parallel and other prepare input/output functions (pytorch#111176)

As part of TP UX improvements, we want to keep our API simple (not easy) so that users get the flexibility to do what they want and avoid a too generic API which tries to solve everything and get things too complicated. We are updating the doc accordingly.

Pull Request resolved: pytorch#111176
Approved by: https://github.com/wanchaol
ghstack dependencies: pytorch#111160, pytorch#111166
yeounoh pushed a commit to yeounoh/pytorch that referenced this pull request Oct 16, 2023
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.

Pull Request resolved: pytorch#111177
Approved by: https://github.com/wanchaol
ghstack dependencies: pytorch#111160, pytorch#111166, pytorch#111176
yeounoh pushed a commit to yeounoh/pytorch that referenced this pull request Oct 16, 2023
@facebook-github-bot facebook-github-bot deleted the gh/fduwjj/109/head branch October 18, 2023 14:25
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk Trigger trunk jobs on your pull request Merged module: dtensor distributed tensor tag release notes: distributed (dtensor) release notes category
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants