-
Notifications
You must be signed in to change notification settings - Fork 767
[MPS] Add support for flatbuffer serialization > 4GB #4574
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[MPS] Add support for flatbuffer serialization > 4GB #4574
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/4574
Note: Links to docs will display an error until the docs builds have been completed. ❌ 2 New FailuresAs of commit 57f5845 with merge base 1b092e9 ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
a77f9cf to
f6a94d4
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great. Thank you for sending the change!
edit: trigger more tests as the end to end stories + llama is in the trunk job and we want to have it cover
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
Thanks @cccclai . Please let me know if I should be triggering these tests, or is this done automatically? |
yeah I've triggered them, and looks clean. |
backends/apple/mps/mps_preprocess.py
Outdated
|
|
||
| import torch | ||
|
|
||
| from backends.apple.mps.serialization.mps_graph_schema import Buffer |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you add this change? The test is failing internally...
| from backends.apple.mps.serialization.mps_graph_schema import Buffer | |
| from executorch.backends.apple.mps.serialization.mps_graph_schema import Buffer |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@cccclai @DenisVieriu97 Do we know why it's not consistent? Make it fail on GitHub is better in this case
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not exactly sure, maybe related to how it's installed
f6a94d4 to
a03dfb5
Compare
|
Thanks! Still have a lint error...https://github.com/pytorch/executorch/actions/runs/10311496755/job/28545193000?pr=4574 |
a03dfb5 to
c669217
Compare
c669217 to
57f5845
Compare
Updated! |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
Add support for serializing tensor weights > 4GB.
This change mirrors #1542 and #1543 for the MPS delegate, which enables serialization of constant weight data outside the flatbuffer blob.
Final raw bytes packed in the PT file have the following format:
cc @cccclai, @larryliu0820, @kimishpatel