New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Intel MKL] Changes in common files to enable MKL Quantized ops with native format #45107
[Intel MKL] Changes in common files to enable MKL Quantized ops with native format #45107
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the PR and I'm so sorry for the delay! The overall PR looks good to me.
I'm a bit hesitant on adding oneDNN-specific parameters (is_*_const
) to the non-MKL-prefixed Quantized ops (which are visible to C++ API users). But I'll leave this part to the API review. (Will tag them after I approve this PR.)
Thank you for reviewing the PR. I have made the requested changed. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the fixes!
For adding an attribute to the ops, I thought these ops are not visible in the API as we mark them as hidden, for e.g. here.
Oh, you're right. Sorry I forgot to check before. :)
However, this PR will still require an API review anyway because it modifies the API golden files (//tensorflow/tools/api/golden/...
). Tagging API review now. (It will be reviewed next week.)
@mahmoud-abuzaina Can you please check @penpornk's comments and keep us posted ? Thanks! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@mahmoud-abuzaina Sorry for the delay! The API review committee would like to see how the added attributes will be used. Can we put these attribute changes in the PRs that actually have the implementation that use them? We can keep other general changes in this PR.
@penpornk no worries. This PR and other related pending PRs don't add any new implementation to use the newly added attributes. Rather, the implementation that uses these attributes has been there for long time. |
Closing for now. I will do some changes and reopen once ready. |
@mahmoud-abuzaina So sorry for my delayed reply and thank you very much for the thorough explanation! I didn't remember that these attributes are already used in the already checked-in kernels. Re: Hidden ops: The API review committee reminded me that op names not prefixed by Edited to add: We synced offline about closing the PR for now. Looking forward to the changes! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm so sorry for the long delay. Thank you again for the PR!
This PR adds the common changes to enable MKL Quantized ops in native format mode. It removes the dependency on the duplicate number of inputs/outputs for quantized ops.
Added by @penpornk for API review:
data_format
is added because MKL ops support both TF's native NHWC format and its own blocked format.is_weight_const
andis_filter_const
are added for caching purposes.