-
-
Notifications
You must be signed in to change notification settings - Fork 608
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Recurrent benchmarks #1871
Recurrent benchmarks #1871
Conversation
One thing I was experimenting with locally was a more structured benchmarking system like https://github.com/JuliaGPU/CUDA.jl/blob/master/perf/runbenchmarks.jl. It would be nice to have a way to visualize results from these smaller-scale performance tests. https://speed.fluxml.ai/ currently has just E2E examples. |
@CarloLucibello and @ToucheSir I'm cool with waiting to add RNN tests until we figure out a better framework if that is preferable! Otherwise, what do we want done in this PR to get this merged? |
be easier to overload behaviour.
I made some modifications. Primarily from @ToucheSir's suggestions. I also modified the bench_utils slightly to make overloading functionality easier. Primarily by adding a pullback function |
Co-authored-by: Brian Chen <ToucheSir@users.noreply.github.com>
Given the conversation in #1855, I think we should have benchmarks in the main repo we can use to iterate on for rnn performance. This touches benchmark_utils to add some of the
Flux.Recur
specific machinery we need. I only added tests forRNN
using both the vector and 3d interfaces, but we can add the other cells once we are happy w/ how these operate.PR Checklist