Skip to content

Fix eval in regression test #1305

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 5 commits into from
Aug 11, 2024
Merged

Conversation

ebsmothers
Copy link
Contributor

@ebsmothers ebsmothers commented Aug 11, 2024

The config name and the results parsing in our regression test job are incorrect.

This is actually a bit awkward to test now that we (a) don't allow creating PRs from a fork, and (b) don't let forks access the S3 bucket containing regression test artifacts.

So for now I've tested it locally, which I guess is better than nothing?

pytest tests/regression_tests/test_llama2_7b.py -m slow_integration_test
...
====== 1 passed, 1 warning in 267.94s (0:04:27) ======

Copy link

pytorch-bot bot commented Aug 11, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchtune/1305

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 56ce980 with merge base 00bbd53 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Aug 11, 2024
@ebsmothers ebsmothers requested a review from joecummings August 11, 2024 04:31
@ebsmothers ebsmothers marked this pull request as draft August 11, 2024 04:40
@ebsmothers ebsmothers changed the title Fix eval config name in regression test [wip] Fix eval config name in regression test Aug 11, 2024
@ebsmothers ebsmothers marked this pull request as ready for review August 11, 2024 05:10
@ebsmothers ebsmothers changed the title [wip] Fix eval config name in regression test Fix eval config name in regression test Aug 11, 2024
@@ -50,7 +50,7 @@ def test_finetune_and_eval(self, tmpdir, caplog, monkeypatch):
runpy.run_path(TUNE_PATH, run_name="__main__")
eval_cmd = f"""
tune run eleuther_eval \
--config eleuther_eval \
--config eleuther_evaluation \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How was this test working before??

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It wasn’t, it’s been failing in CI for a while now. But it doesn’t run on PRs or anything, only nightly

@SalmanMohammadi
Copy link
Contributor

This is actually a bit awkward to test now that we (a) don't allow creating PRs from a fork, and (b) don't let forks access the S3 bucket containing regression test artifacts.

I missed this, but we're no longer running integration tests in our CI?

@ebsmothers
Copy link
Contributor Author

This is actually a bit awkward to test now that we (a) don't allow creating PRs from a fork, and (b) don't let forks access the S3 bucket containing regression test artifacts.

I missed this, but we're no longer running integration tests in our CI?

So we actually have two types of integration tests: recipe tests and regression tests. Recipe tests always run on PRs but only use small checkpoints. Regression tests use the full-size model and run nightly. Currently this is the only regression test we have, but we’ve been wanting to add more and just haven’t had time (e.g. it’d be nice if we could test memory or perf of some of our models too).

@SalmanMohammadi
Copy link
Contributor

I'll raise an issue : )

@ebsmothers ebsmothers changed the title Fix eval config name in regression test Fix eval in regression test Aug 11, 2024
@ebsmothers ebsmothers merged commit cc988f2 into pytorch:main Aug 11, 2024
20 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants