-
Notifications
You must be signed in to change notification settings - Fork 25.7k
[quant][embedding qat] Add eager QAT test for EmbeddingBag+Linear model #66334
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
[ghstack-poisoned]
CI Flow Status⚛️ CI FlowRuleset - Version:
You can add a comment to the PR and tag @pytorchbot with the following commands: # ciflow rerun, "ciflow/default" will always be added automatically
@pytorchbot ciflow rerun
# ciflow rerun with additional labels "-l <ciflow/label_name>", which is equivalent to adding these labels manually and trigger the rerun
@pytorchbot ciflow rerun -l ciflow/scheduled -l ciflow/slowFor more information, please take a look at the CI Flow Wiki. |
🔗 Helpful links
💊 CI failures summary and remediationsAs of commit 02b5595 (more details on the Dr. CI page): 💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.Please report bugs/suggestions to the (internal) Dr. CI Users group. |
…+Linear model" [ghstack-poisoned]
…+Linear model" [ghstack-poisoned]
…+Linear model" [ghstack-poisoned]
|
@b-koopman has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
|
|
||
| def forward(self, input: torch.Tensor, offsets: Optional[torch.Tensor] = None, | ||
| per_sample_weights: Optional[torch.Tensor] = None): | ||
| x = self.emb(input, offsets, per_sample_weights) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
isn't the input to nn.quantized.EmbeddingBag supposed to be quantized, so sholdn't this be inside the QuantStub()?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Input indices shouldn't be quantized, only the weights, so this shouldn't be wrapped in quant/dequant as you would expect for a linear layer.
…+Linear model" Differential Revision: [D31618283](https://our.internmc.facebook.com/intern/diff/D31618283) [ghstack-poisoned]
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good but I'd make sure to clear that failing xenial test before trying to land or at least figure out if its okay to ignore.
…+Linear model" Differential Revision: [D31618283](https://our.internmc.facebook.com/intern/diff/D31618283) [ghstack-poisoned]
|
@b-koopman has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Stack from ghstack:
Differential Revision: D31618283