-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error for unsupported precision types with ModelParallelStrategy #19902
Conversation
for more information, see https://pre-commit.ci
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
A comment on making the error message simpler to action
for more information, see https://pre-commit.ci
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## master #19902 +/- ##
=========================================
- Coverage 84% 59% -25%
=========================================
Files 426 421 -5
Lines 35247 35156 -91
=========================================
- Hits 29621 20752 -8869
- Misses 5626 14404 +8778 |
What does this PR do?
Adds error handling if user passes
precision="16-mixed"
. This would require a sharded gradscaler to be configured, but we opted for not supporting this at the moment.precision="bf16-mixed"
(bfloat) is still supported, as it doesn't require a grad scaler. Other precisions such as bitsandbytes are also not allowed (they don't compose with this strategy) by design.cc @Borda @carmocca @justusschock @awaelchli