-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tensorflow optimizers approach NaN #37
Comments
Thank you for reporting this @phgilde , we'll look into it. |
same issue. running on an AMD Vega 64 and Ryzen 3700X |
It looks like the issue has already been addressed internally. I can repro the NaN convergence on the latest pypi build, but not with our latest internal build. I'll let you know once we release a new wheel so you can try it out! |
Is there any ETA for the new wheel? Thank you very much. |
We just released tensorflow-directml 1.15.3.dev200911, which should contain the fixes for the NaN errors that you were seeing. You can try it out and tell us how it goes! Also, since we have now open-sourced our fork, new tensorflow-directml issues should be opened over here. |
This model/training loop approaches an error of NaN after a couple of iterations:
After a couple of training loops, the loss is NaN instead of a float
System: Intel i5-7200U with Intel HD Graphics 620
The text was updated successfully, but these errors were encountered: