-
Notifications
You must be signed in to change notification settings - Fork 25.7k
[inductor][addmm] incorporate into new get_mm_configs properly #161534
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: gh/coconutruben/55/base
Are you sure you want to change the base?
Conversation
\# why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic \# what - expand KernelInputs to also store views of specific nodes, by names - use that view (inp, the unexpanded version) in the heuristics to adjust it depending on whether we're in max-autotune or not \# testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` [ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/161534
Note: Links to docs will display an error until the docs builds have been completed. ❌ 5 New FailuresAs of commit 66e7441 with merge base d25c35d ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - expand KernelInputs to also store views of specific nodes, by names - use that view (inp, the unexpanded version) in the heuristics to adjust it depending on whether we're in max-autotune or not # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - expand KernelInputs to also store views of specific nodes, by names - use that view (inp, the unexpanded version) in the heuristics to adjust it depending on whether we're in max-autotune or not # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - expand KernelInputs to also store views of specific nodes, by names - use that view (inp, the unexpanded version) in the heuristics to adjust it depending on whether we're in max-autotune or not # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - remove the view from inp_expanded when running not in max-autotune # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
2 similar comments
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - remove the view from inp_expanded when running not in max-autotune # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Failing tests?
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
@coconutruben has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
\# why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic \# what - remove the view from inp when not in max-autotune for addmm aten \# testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` ghstack-source-id: 4399549 Pull Request resolved: #161534
…erly" # why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic # what - pass unexpanded bias (inp) - let template (heuristics) that it to be expanded (ATen in not max-autotune, Triton always) expand it # testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy chenyang78 kadeng muchulee8 amjames chauhang aakhundov Differential Revision: [D81520581](https://our.internmc.facebook.com/intern/diff/D81520581) [ghstack-poisoned]
\# why - addmm aten running with an expanded version of bias vs the regular bias sometimes causes numerics differences - to avoid this for now, we make addmm aten use inp vs inp_expanded depending on if we're in max-autotune or not, matching the previous logic \# what - remove the view from inp when not in max-autotune for addmm aten \# testing ``` python3 -bb -m pytest test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_debug_printer_codegen_cpu ``` ghstack-source-id: 208c906 Pull Request resolved: #161534
Stack from ghstack (oldest at bottom):
why
bias sometimes causes numerics differences
depending on if we're in max-autotune or not, matching the previous
logic
what
testing
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @chenyang78 @kadeng @muchulee8 @amjames @chauhang @aakhundov
Differential Revision: D81520581