-
Notifications
You must be signed in to change notification settings - Fork 536
[ExecuTorch][Weight Sharing][XNNPACK] load named data map data for xnnpack #9152
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…npack If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/) [ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/9152
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 935f105 with merge base 630d0cc ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This pull request was exported from Phabricator. Differential Revision: D70315209 |
…data for xnnpack" If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/) [ghstack-poisoned]
This pull request was exported from Phabricator. Differential Revision: D70315209 |
…data for xnnpack" If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/) [ghstack-poisoned]
This pull request was exported from Phabricator. Differential Revision: D70315209 |
…data for xnnpack" If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/) [ghstack-poisoned]
This pull request was exported from Phabricator. Differential Revision: D70315209 |
…data for xnnpack" If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/) [ghstack-poisoned]
This pull request was exported from Phabricator. Differential Revision: D70315209 |
…npack (#9294) This PR was created by the merge bot to help merge the original PR into the main branch. ghstack PR number: #9152 by @mcr229 ^ Please use this as the source of truth for the PR details, comments, and reviews ghstack PR base: https://github.com/pytorch/executorch/tree/gh/mcr229/8/base ghstack PR head: https://github.com/pytorch/executorch/tree/gh/mcr229/8/head Merge bot PR base: https://github.com/pytorch/executorch/tree/gh/mcr229/7/orig Merge bot PR head: https://github.com/pytorch/executorch/tree/gh/mcr229/8/orig @diff-train-skip-merge --------- Co-authored-by: Max Ren <maxren@meta.com>
…npack Pull Request resolved: #9152 If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload. Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble. Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights. ghstack-source-id: 271732048 @exported-using-ghexport Differential Revision: [D70315209](https://our.internmc.facebook.com/intern/diff/D70315209/)
…npack (pytorch#9294) This PR was created by the merge bot to help merge the original PR into the main branch. ghstack PR number: pytorch#9152 by @mcr229 ^ Please use this as the source of truth for the PR details, comments, and reviews ghstack PR base: https://github.com/pytorch/executorch/tree/gh/mcr229/8/base ghstack PR head: https://github.com/pytorch/executorch/tree/gh/mcr229/8/head Merge bot PR base: https://github.com/pytorch/executorch/tree/gh/mcr229/7/orig Merge bot PR head: https://github.com/pytorch/executorch/tree/gh/mcr229/8/orig @diff-train-skip-merge --------- Co-authored-by: Max Ren <maxren@meta.com>
Stack from ghstack (oldest at bottom):
If data is serialized into the NamedDataMap, then we overload getConstantDataPtr to retrieve the data from the named data map. This should be done in a Backwards Compatible way. Meaning if no data is serialized into the named data map, then we are still loading the data from the flatbuffer payload.
Since the runtime change here is being made before the AoT changes, All CI on this diff by itself should test that the changes made here are backwards compatitble.
Note: We do not resolve Runtime Memory usage at this point. WeightCache will be implemented in the next diff. Meaning If we load via the same key across different methods, we still pack twice and allocate two instances for the packed weights.
Differential Revision: D70315209