-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] DeepSpeed non-deterministic inference with HF GPT2 when replace_with_kernel_inject=True
#2243
Comments
Okay, I verify that by changing the |
Hi @trianxy, I'm sorry for the lack of updates on this, but with latest master (should be released as 0.7.5 in the next few days) I believe the issue you're observing here is fixed. Would you mind testing this on your end to verify if this is true? Thanks! |
Thank you @cmikeh2 for coming back to me on that. I think the above issue can be closed, because it is fixed in versions Does the fact, that it works already in I am happy to do additional tests. |
|
Describe the bug
#1950 describes a bug by which running inference twice on the same input leads to different outputs. It was supposedly fixed in version
0.6.5
, but I am encountering a similar bug (for Huggingface's GPT2, on an NVidia A10G) in every deepspeed version after including0.6.3
when running long sequences. My current fix is to use version0.6.1
.Note: When running too short a sequence this bug does not appear. When running too long a sequence, I am rather seeing another open bug (#2062) which prohibits inference.
Perhaps related bug: #2229
To Reproduce
Expected behavior
I expected to see the same output each time, i.e.
ds_report output
System info (please complete the following information):
Launcher context
inside a Python notebook
The text was updated successfully, but these errors were encountered: