-
Notifications
You must be signed in to change notification settings - Fork 25.7k
Closed
Labels
module: dtensordistributed tensor tagdistributed tensor tagoncall: distributedAdd this issue/PR to distributed oncall triage queueAdd this issue/PR to distributed oncall triage queue
Description
🚀 The feature, motivation and pitch
Would be great to have a sharding strategy for aten.view.dtype. Currently, when doing .view(dtype) on a DTensor, you get the following error:
Operator aten.view.dtype does not have a sharding strategy registered.
This has caused issues for FSDP2 + stochastic rounding for me. See: pytorch/ao#1505
Alternatives
No response
Additional context
No response
cc @H-Huang @awgu @kwen2501 @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o @tianyu-l @XilunWu
Metadata
Metadata
Assignees
Labels
module: dtensordistributed tensor tagdistributed tensor tagoncall: distributedAdd this issue/PR to distributed oncall triage queueAdd this issue/PR to distributed oncall triage queue