-
Notifications
You must be signed in to change notification settings - Fork 25.6k
[dtensor] move pad/unpad_tensor to separate utils #124871
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument [ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/124871
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit ba282a6 with merge base e592a60 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Looks like need to migrate
|
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument cc mrshenli pritamdamania87 zhaojuanmao satgera rohan-varma gqchen aazzolini osalpekar jiayisuse H-Huang kwen2501 awgu penguinwu fegin XilunWu fduwjj wz337 tianyu-l wconstab yf225 chauhang d4l3k [ghstack-poisoned]
good catch! updated |
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Merge failedReason: 4 jobs have failed, first few of them are: linux-binary-libtorch-cxx11-abi, trunk, linux-binary-manywheel, linux-binary-libtorch-pre-cxx11 Details for Dev Infra teamRaised by workflow job |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Merge failedReason: 4 jobs have failed, first few of them are: linux-binary-libtorch-cxx11-abi, trunk, linux-binary-manywheel, linux-binary-libtorch-pre-cxx11 Details for Dev Infra teamRaised by workflow job |
@pytorchbot merge -i |
@pytorchbot revert -m "Broke internal tests, see D56587991 for more details" -c ignoredsignal |
@pytorchbot successfully started a revert job. Check the current status here. |
This reverts commit 0b0eea2. Reverted #124871 on behalf of https://github.com/jeanschmidt due to Broke internal tests, see D56587991 for more details ([comment](#124871 (comment)))
@wanchaol your PR has been successfully reverted. |
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument Pull Request resolved: pytorch#124871 Approved by: https://github.com/awgu, https://github.com/wz337
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument cc mrshenli pritamdamania87 zhaojuanmao satgera rohan-varma gqchen aazzolini osalpekar jiayisuse H-Huang kwen2501 awgu penguinwu fegin XilunWu fduwjj wz337 tianyu-l wconstab yf225 chauhang d4l3k [ghstack-poisoned]
as titled, we implement a dedicated communication op to allow efficient sharding dimension change using alltoall, to replace our previous allgather + local chunk Pull Request resolved: #124872 Approved by: https://github.com/XilunWu, https://github.com/yifuwang ghstack dependencies: #124871
as titled, as we have a dedicated comm op, this is not needed anymore Pull Request resolved: #124879 Approved by: https://github.com/XilunWu, https://github.com/wz337 ghstack dependencies: #124871, #124872
as titled, we implement a dedicated communication op to allow efficient sharding dimension change using alltoall, to replace our previous allgather + local chunk Pull Request resolved: #124872 Approved by: https://github.com/XilunWu, https://github.com/yifuwang ghstack dependencies: #124871
as titled, as we have a dedicated comm op, this is not needed anymore Pull Request resolved: #124879 Approved by: https://github.com/XilunWu, https://github.com/wz337 ghstack dependencies: #124871, #124872
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument Pull Request resolved: #124871 Approved by: https://github.com/awgu, https://github.com/wz337
This reverts commit 0b0eea2. Reverted #124871 on behalf of https://github.com/jeanschmidt due to Broke internal tests, see D56587991 for more details ([comment](#124871 (comment)))
as titled, 1. pad/unpad is a general util not specific to the Shard placement, 2. for the propose of the next PR, move these two out of Shard placement itself, and give additional pad_dim argument Pull Request resolved: pytorch#124871 Approved by: https://github.com/awgu, https://github.com/wz337, https://github.com/XilunWu
as titled, we implement a dedicated communication op to allow efficient sharding dimension change using alltoall, to replace our previous allgather + local chunk Pull Request resolved: #124872 Approved by: https://github.com/XilunWu, https://github.com/yifuwang ghstack dependencies: #124871
as titled, as we have a dedicated comm op, this is not needed anymore Pull Request resolved: pytorch#124879 Approved by: https://github.com/XilunWu, https://github.com/wz337 ghstack dependencies: pytorch#124871, pytorch#124872
as titled, we implement a dedicated communication op to allow efficient sharding dimension change using alltoall, to replace our previous allgather + local chunk Pull Request resolved: pytorch#124872 Approved by: https://github.com/XilunWu, https://github.com/yifuwang ghstack dependencies: pytorch#124871
as titled, as we have a dedicated comm op, this is not needed anymore Pull Request resolved: #124879 Approved by: https://github.com/XilunWu, https://github.com/wz337 ghstack dependencies: #124871, #124872
Stack from ghstack (oldest at bottom):
as titled, 1. pad/unpad is a general util not specific to the Shard
placement, 2. for the propose of the next PR, move these two out of Shard
placement itself, and give additional pad_dim argument
cc @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @H-Huang @kwen2501 @awgu @penguinwu @fegin @XilunWu @fduwjj @wz337 @tianyu-l @wconstab @yf225 @chauhang @d4l3k