-
Notifications
You must be signed in to change notification settings - Fork 668
Fix Android app S3 path with tiktoken #4289
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/4289
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New FailureAs of commit 13f29f0 with merge base 1d7d71d ( NEW FAILURE - The following job has failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
@huydhn has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
Thanks for the quick fix @huydhn. I think it's better to temporarily disable/remove it from this android job. This android job runs on the PR and should only be used for quick validation of the android demo app build and run with a tiny model. It also runs in the GitHub action to prompt commit from main to stable, the instability would affect the stable branch health and release as we would prefer doing branch-cut off stable if possible. So the better place is to run the benchmarking job in a non-critical path. I have #4287 to remove/disable the uploading and benchmarking portion off the android job |
# mentions that tiktoken is only for Llama3. So, we can export it later in another archive | ||
# like https://ossci-assets.s3.amazonaws.com/executorch-android-llama2-7b.zip when this is | ||
# updated to run Llama3 | ||
tiktoken: [OFF] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the reason of #4203 updating this to include tiktoken tokenizer is to ensure the demo app can build for Llama3. So as mentioned that this job should focus on validating different flavors of android demo app builds. If we remove tiktoken here, there will be a tcc regression.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, this covers only llama2 without tiktoken atm. To cover llama3, for example, we would need to zip the exported model and the tokenizer it uses and upload the archive onto S3, i.e. https://github.com/pytorch/executorch/blob/main/.github/workflows/android.yml#L122. Then the archive can be passed to the job as its extra-data
parameter to be copied to the device.
After #4203, the app is now uploaded to
artifact/tiktoken_${{ matrix.tiktoken }}
instead withtiktoken
can be eitherON
orOFF
. https://github.com/pytorch/executorch/blob/main/examples/demo-apps/android/LlamaDemo/README.md#alternative-2-build-from-local-machine mentions that tiktoken is only for Llama3. So, we can export it later in another archive like https://ossci-assets.s3.amazonaws.com/executorch-android-llama2-7b.zip when this is updated to run Llama3.Note that the test is still unstable because the app runs OOM on S22 (#3344)