Skip to content

Allow each model to choose its own runner #277

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

huydhn
Copy link
Contributor

@huydhn huydhn commented Sep 12, 2023

It's a good idea to allow each model to choose its own runner. This also help unblock #212 until we know the root cause of why it runs OOM.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Sep 12, 2023
@huydhn huydhn requested a review from kirklandsign September 12, 2023 02:40
@huydhn huydhn marked this pull request as ready for review September 12, 2023 02:40
@facebook-github-bot
Copy link
Contributor

@huydhn has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator.

@facebook-github-bot
Copy link
Contributor

@huydhn merged this pull request in 89f3e39.

facebook-github-bot pushed a commit that referenced this pull request Sep 13, 2023
Summary:
I mistakenly replace this in #277.  The runner is not set here, and causes syntax error, for example https://github.com/pytorch/executorch/actions/runs/6167285165.  The workflow still works because the default `linux.2xlarge` is used (same runner type)

### Testing

The annotation error is gone now https://github.com/pytorch/executorch/actions/runs/6175961826

Pull Request resolved: #313

Test Plan: https://github.com/pytorch/executorch/actions/runs/6175961826

Reviewed By: mergennachin

Differential Revision: D49236002

Pulled By: huydhn

fbshipit-source-id: bd5282f582d23427fd1d7622ed3b2ee2e6ad93cb
Gasoonjia pushed a commit that referenced this pull request Jul 30, 2024
Add special case for zero-temperature sampling. For stories15M on my devserver, seems to improve tokens/sec as follows:

before: 189, 180, 166
after: 264, 285, 285
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants