-
Notifications
You must be signed in to change notification settings - Fork 90
Remove conv weight scale reciprocal #222
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove conv weight scale reciprocal #222
Conversation
|
Hi @jgong5 @Xia-Weiwen, could you help to take a look of this PR? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. This depends on onednn 3.x right, otherwise would result in incorrectness, right?
Yes, we need to merge PR: pytorch/pytorch#105996 firstly before update PyTorch IDeep to this commit. |
|
Hi @yanbing-j, could you help to merge it? |
…tion in QConv PT2E" **Summary** After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance. Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed. cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
**Summary** After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance. Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed. cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
…tion in QConv PT2E" **Summary** After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance. Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed. cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
**Summary** After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance. Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed. cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
…ndant QConv weight scale reciprocal calculation" **Summary** Upgrade IDeep, the only diff IDeep change is this IDeep PR: intel/ideep#222 which has done 2 things: - Remove the redundant QConv weight scale reciprocal calculation. - Pump IDEEP_VERSION_REVISION version from 0 to 1. So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch. cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
…ht scale reciprocal calculation" **Summary** Upgrade IDeep, the only diff IDeep change is this IDeep PR: intel/ideep#222 which has done 2 things: - Remove the redundant QConv weight scale reciprocal calculation. - Pump IDEEP_VERSION_REVISION version from 0 to 1. So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch. cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
…ndant QConv weight scale reciprocal calculation" **Summary** Upgrade IDeep which includes 2 IDeep change as IDeep PR: intel/ideep#222 and intel/ideep#223 - For IDeep PR: intel/ideep#222 which has done 2 things: - Remove the redundant QConv weight scale reciprocal calculation. - Pump IDEEP_VERSION_REVISION version from 0 to 1. So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch. - For IDeep PR: intel/ideep#223 which includes AArch64 specific changes with the oneDNN 3.1.1 upgrade. cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
…ht scale reciprocal calculation" **Summary** Upgrade IDeep which includes 2 IDeep change as IDeep PR: intel/ideep#222 and intel/ideep#223 - For IDeep PR: intel/ideep#222 which has done 2 things: - Remove the redundant QConv weight scale reciprocal calculation. - Pump IDEEP_VERSION_REVISION version from 0 to 1. So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch. - For IDeep PR: intel/ideep#223 which includes AArch64 specific changes with the oneDNN 3.1.1 upgrade. cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen [ghstack-poisoned]
After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.