You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I know that Zero is available in Fairscale, but since PyTorch now has its own implementation at torch.distributed.optim.ZeroRedundancyOptimizer, I think it would make sense to support it.
Pitch
A new strategy called e.g. DDPZero that would support DDP with torch's ZeroRedundancyOptimizer, along with support for overlapping ZeroRedundancyOptimizer with DDP ("overlap_with_ddp") and the related DDP comms hooks.
Alternatives
As mentioned above, Fairscale implements Zero. However, this requires taking on an additional dependency. And since pytorch has a native implementation, it could make sense to support that in lightning.
The text was updated successfully, but these errors were encountered:
馃殌 Feature
Implement a DDP-ZeroRedundancyOptimizer strategy.
Motivation
I know that Zero is available in Fairscale, but since PyTorch now has its own implementation at
torch.distributed.optim.ZeroRedundancyOptimizer
, I think it would make sense to support it.Pitch
A new
strategy
called e.g.DDPZero
that would supportDDP
with torch'sZeroRedundancyOptimizer
, along with support for overlapping ZeroRedundancyOptimizer with DDP ("overlap_with_ddp
") and the related DDP comms hooks.Alternatives
As mentioned above, Fairscale implements Zero. However, this requires taking on an additional dependency. And since pytorch has a native implementation, it could make sense to support that in lightning.
The text was updated successfully, but these errors were encountered: