Skip to content

Conversation

kimishpatel
Copy link
Contributor

Summary: Dequantize gemm when doing prefill like op, else use custom kernel

Reviewed By: metascroy

Differential Revision: D71833065

Summary: Dequantize gemm when doing prefill like op, else use custom kernel

Reviewed By: metascroy

Differential Revision: D71833065
Copy link

pytorch-bot bot commented Apr 11, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/10108

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 056a4b7 with merge base a073668 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Apr 11, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71833065

@kimishpatel kimishpatel added the release notes: examples Changes to any of our example LLMs integrations, such as Llama3 and Llava label Apr 11, 2025
@facebook-github-bot facebook-github-bot merged commit 4022ff1 into pytorch:main Apr 14, 2025
83 of 87 checks passed
keyprocedure pushed a commit to keyprocedure/executorch that referenced this pull request Apr 21, 2025
Differential Revision: D71833065

Pull Request resolved: pytorch#10108
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported release notes: examples Changes to any of our example LLMs integrations, such as Llama3 and Llava

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants