-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Shell script to collect benchmarks for multiple versions #15144
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
One problem that I sometimes encounter is that cargo decides to use |
Sorry @logan-keede -- I am very excited to try this one out, but I ran otu of time today |
No problem at all! I look forward to your feedback whenever you have the time. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @logan-keede that is pretty neat, appreciate if you document the script and its usage in benchmarks/README.md
Just noticed that benchmarks/README.md is not included in prettier CI check, is that intended? |
I don't think it is intended And I am sorry I haven't had time to test out this PR yet |
I pushed a commit to add benchmarks to prettier check list. No problem, I understand it might be a bit to lengthy and resource intensive to test. |
Thank you for this work @logan-keede -- I am sorry for the very long delay. It seems @saraghds is also working on this script so I am going to try and help it along. I merged this branch up from main and added some more comments. I also would like to make the versions it checks a bit more configurable -- I will test and work on that over the day |
I want this script to be useable by more people so I think it is important to document it a bit more I am also going to investigate potentially overriding the list of git commits to run rather than assuming what versions. |
Sorry for the long delay in feedback -- I have been working on other things. What I really want is a way to track DataFusion performance over time during development (so not released version numbers) so that as new changes are added to the code we can see how the overall trend is doing. Given the initial state of this PR, the best they can do is run once a month after a release is done. So since they don't really satisfy the need yet (being able to get a handle on ongoing performane) I haven't spent a lot of time reviewing / working them. BTW here is the kind of thing I want to see / create: ![]() |
Which issue does this PR close?
Rationale for this change
Originally posted by @alamb in #5504
What changes are included in this PR?
Just a simple script to collect benchmarks for last 5 releases
Are these changes tested?
Yes, by running the script on my laptop.
using
sh collect_bench.sh tpch
.Are there any user-facing changes?
Nope.