You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Either through torch amp or via a postfix .half() cast (not too nice), make sure that we don't reduce fp32 grads when they were actually computed on fp16..
Motivation
The gradient size are double what they should be if torch AMP is being used, which can be a bottleneck depending on internode coms
Pitch
free speed
Alternatives
Not doing anything, this does not affect correctness
Additional context
The text was updated successfully, but these errors were encountered:
馃殌 Feature
Either through torch amp or via a postfix .half() cast (not too nice), make sure that we don't reduce fp32 grads when they were actually computed on fp16..
Motivation
The gradient size are double what they should be if torch AMP is being used, which can be a bottleneck depending on internode coms
Pitch
free speed
Alternatives
Not doing anything, this does not affect correctness
Additional context
The text was updated successfully, but these errors were encountered: