- 
                Notifications
    
You must be signed in to change notification settings  - Fork 559
 
Revert "Fix model initialization. (#6076)" #6134
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
| 
           Sorry I didn't catch up with recent PRs. Do we already support automatically match dynamo backend   | 
    
| 
           We should support openxla_eval + eval, openxla + eval, and openxla + train  | 
    
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the clarification. LGTM! Can you also resolve the conflict?
c67100e    to
    6e61d8d      
    Compare
  
    This reverts commit 7786d5d. `python benchmarks/experiment_runner.py --dynamo=openxla --dynamo=openxla_eval --dynamo=inductor --xla=PJRT --xla=None --test=eval --test=train --suite-name=torchbench --accelerator=cuda --filter=^alexnet$ --no-resume` fails for openxla benchmarks.
6e61d8d    to
    833ec06      
    Compare
  
    I discovered some models from the suite do not have the precision set so instead of failing the script we just log the case, and use the default precision, as no additional machinery should run for the Inductor anyway. Additionally I wrapped the exceptions with the ValueError so the logging message will not pollute with info about str not inheriting from Exception class. ecg@, note that needs to be hooked "somewhere". Not sure where, as there was a revert in #6134.
I discovered some models from the suite do not have the default precision set so instead of failing the script we just log the case, and do nothing, as no additional machinery should run for the Inductor anyway. Additionally I wrapped the exceptions with the ValueError so the logging message will not pollute with info about str not inheriting from Exception class. @cota , note that needs to be hooked "somewhere". Not sure where, as there was a revert in #6134, but in general it can be done prior to moving the model to the device safely.
I discovered some models from the suite do not have the default precision set so instead of failing the script we just log the case, and do nothing, as no additional machinery should run for the Inductor anyway. Additionally I wrapped the exceptions with the ValueError so the logging message will not pollute with info about str not inheriting from Exception class. @cota , note that needs to be hooked "somewhere". Not sure where, as there was a revert in pytorch#6134, but in general it can be done prior to moving the model to the device safely.
I discovered some models from the suite do not have the default precision set so instead of failing the script we just log the case, and do nothing, as no additional machinery should run for the Inductor anyway. Additionally I wrapped the exceptions with the ValueError so the logging message will not pollute with info about str not inheriting from Exception class. @cota , note that needs to be hooked "somewhere". Not sure where, as there was a revert in #6134, but in general it can be done prior to moving the model to the device safely.
I discovered some models from the suite do not have the default precision set so instead of failing the script we just log the case, and do nothing, as no additional machinery should run for the Inductor anyway. Additionally I wrapped the exceptions with the ValueError so the logging message will not pollute with info about str not inheriting from Exception class. @cota , note that needs to be hooked "somewhere". Not sure where, as there was a revert in #6134, but in general it can be done prior to moving the model to the device safely.
This reverts commit 7786d5d.
python benchmarks/experiment_runner.py --dynamo=openxla --dynamo=openxla_eval --dynamo=inductor --xla=PJRT --xla=None --test=eval --test=train --suite-name=torchbench --accelerator=cuda --filter=^alexnet$ --no-resumefails for openxla benchmarks.