Skip to content

Conversation

@leslie-fang-intel
Copy link
Contributor

After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.

@leslie-fang-intel
Copy link
Contributor Author

Hi @jgong5 @Xia-Weiwen, could you help to take a look of this PR?

Copy link

@jgong5 jgong5 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM. This depends on onednn 3.x right, otherwise would result in incorrectness, right?

@leslie-fang-intel
Copy link
Contributor Author

LGTM. This depends on onednn 3.x right, otherwise would result in incorrectness, right?

Yes, we need to merge PR: pytorch/pytorch#105996 firstly before update PyTorch IDeep to this commit.

@leslie-fang-intel
Copy link
Contributor Author

Hi @yanbing-j, could you help to merge it?

leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 18, 2023
…tion in QConv PT2E"


**Summary**
After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.
Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed.

cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 18, 2023
**Summary**
After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.
Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed.

cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
@yanbing-j yanbing-j merged commit c90e52c into intel:ideep_pytorch Aug 19, 2023
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 19, 2023
…tion in QConv PT2E"


**Summary**
After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.
Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed.

cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 19, 2023
**Summary**
After oneDNN 3.1 upgrade, we don't need do the weight scale reciprocal calculation. So, remove the duplicated reciprocal calculation to optimize QConv performance.
Change to new IDeep Commit after IDeep PR: intel/ideep#222 landed.

cc jgong5 mingfeima XiaobingSuper sanchitintel ashokei jingxu10 gujinghui PenghuiCheng jianyuh min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 21, 2023
…ndant QConv weight scale reciprocal calculation"


**Summary**
Upgrade IDeep, the only diff IDeep change is this IDeep PR: intel/ideep#222 which has done 2 things:

- Remove the redundant QConv weight scale reciprocal calculation.
- Pump IDEEP_VERSION_REVISION version from 0 to 1.

So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch.


cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 21, 2023
…ht scale reciprocal calculation"


**Summary**
Upgrade IDeep, the only diff IDeep change is this IDeep PR: intel/ideep#222 which has done 2 things:

- Remove the redundant QConv weight scale reciprocal calculation.
- Pump IDEEP_VERSION_REVISION version from 0 to 1.

So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch.


cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 23, 2023
…ndant QConv weight scale reciprocal calculation"


**Summary**
Upgrade IDeep which includes 2 IDeep change as IDeep PR: intel/ideep#222 and intel/ideep#223

- For IDeep PR: intel/ideep#222 which has done 2 things:

  - Remove the redundant QConv weight scale reciprocal calculation.
  - Pump IDEEP_VERSION_REVISION version from 0 to 1.
  
  So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch.

- For IDeep PR: intel/ideep#223 which includes AArch64 specific changes with the oneDNN 3.1.1 upgrade.


cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
leslie-fang-intel added a commit to pytorch/pytorch that referenced this pull request Aug 23, 2023
…ht scale reciprocal calculation"


**Summary**
Upgrade IDeep which includes 2 IDeep change as IDeep PR: intel/ideep#222 and intel/ideep#223

- For IDeep PR: intel/ideep#222 which has done 2 things:

  - Remove the redundant QConv weight scale reciprocal calculation.
  - Pump IDEEP_VERSION_REVISION version from 0 to 1.
  
  So only QConv related calculation will be impacted and we already use IDeep version API in #105996 to make the corresponding change in PyTorch.

- For IDeep PR: intel/ideep#223 which includes AArch64 specific changes with the oneDNN 3.1.1 upgrade.


cc gujinghui PenghuiCheng XiaobingSuper jianyuh jgong5 mingfeima sanchitintel ashokei jingxu10 min-jean-cho yanbing-j Guobing-Chen Xia-Weiwen

[ghstack-poisoned]
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants