Skip to content

Experiment with private rooted Pixel 3 devices #10192

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 7 commits into from
Apr 15, 2025
Merged

Conversation

huydhn
Copy link
Contributor

@huydhn huydhn commented Apr 15, 2025

I'm creating a new experimental workflow to run benchmarks on private rooted Pixel 3 devices. We have only 2 devices there though, so I think let's reduce the number of model we want to cover for now.

Testing

https://github.com/pytorch/executorch/actions/runs/14465442714

Copy link

pytorch-bot bot commented Apr 15, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/10192

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit e0b834c with merge base 4559a61 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Apr 15, 2025
@huydhn huydhn added topic: not user facing and removed CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. labels Apr 15, 2025
@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Apr 15, 2025
@huydhn huydhn temporarily deployed to upload-benchmark-results April 15, 2025 10:36 — with GitHub Actions Inactive
@huydhn huydhn temporarily deployed to upload-benchmark-results April 15, 2025 17:02 — with GitHub Actions Inactive
id-token: write
contents: read
with:
models: ${{ inputs.models }}
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@guangy10 @kirklandsign Appreciate if you could pass me the list of models and benchmark configs that you want to cover here. We have only 2 devices there though, so I think let's reduce the number of model we want to cover for now.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's pick one non-genAI model (mv3, xnnpack_q8) and one genAI model (llama3.2-1b, spinquant & qlora)

mv3,meta-llama/Llama-3.2-1B-Instruct-SpinQuant_INT4_EO8,meta-llama/Llama-3.2-1B-Instruct-QLORA_INT4_EO8

@huydhn huydhn marked this pull request as ready for review April 15, 2025 18:18
contents: read
with:
models: ${{ inputs.models }}
devices: google_pixel_3_private_rooted
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Unfortunately we don't have data points of using the same type of device on public pool, can't do any comparison, so we would mainly rely on the new data on the private & rooted device and compare those across different runs.

Copy link
Contributor

@guangy10 guangy10 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I recall @kirklandsign mentioned that the metrics instability is frequently seen on nightly batched jobs, on-demand with only 1-2 models are running much stable.

@huydhn In this new schedule, and to gather more data points, we may want to run this workflow more often (every 2 hours maybe? Should have enough time to cooldown from previous run)

@huydhn Do we need to merge this PR? I thought we can reuse the existing workflow instead of creating a new fork

@huydhn
Copy link
Contributor Author

huydhn commented Apr 15, 2025

@huydhn Do we need to merge this PR? I thought we can reuse the existing workflow instead of creating a new fork

This uses the existing workflow but gives us the flexibility to use a different schedule and default inputs

@huydhn huydhn merged commit 683869d into main Apr 15, 2025
5 checks passed
@huydhn huydhn deleted the experiment-private-devices branch April 15, 2025 21:36
keyprocedure pushed a commit to keyprocedure/executorch that referenced this pull request Apr 21, 2025
I'm creating a new experimental workflow to run benchmarks on private
rooted Pixel 3 devices. We have only 2 devices there though, so I think
let's reduce the number of model we want to cover for now.

### Testing

https://github.com/pytorch/executorch/actions/runs/14465442714
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. topic: not user facing
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants