Skip to content

Conversation

YifanShenSZ
Copy link
Collaborator

4-bit groupwise weight-only quantization is well supported in Core ML. Since torchao offers the same quantization, let's use torchao to quantize llama then delegate to Core ML.

Copy link

pytorch-bot bot commented Aug 23, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/4866

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 30 Pending

As of commit ac94cd9 with merge base d3da92d (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Aug 23, 2024
@digantdesai
Copy link
Contributor

cc @kimishpatel

@digantdesai digantdesai added triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module module: coreml Issues related to Apple's Core ML delegation and code under backends/apple/coreml/ labels Aug 23, 2024
@YifanShenSZ
Copy link
Collaborator Author

Turns out, we may prefer Core ML quantization at this moment. Landed the alternative in #5228 Locally confirmed performance gain & accuracy preservation

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. module: coreml Issues related to Apple's Core ML delegation and code under backends/apple/coreml/ triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants