-
Notifications
You must be signed in to change notification settings - Fork 221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
updating docs to include matrix-vector multiply example #918
base: master
Are you sure you want to change the base?
Conversation
Codecov Report
@@ Coverage Diff @@
## master #918 +/- ##
==========================================
- Coverage 77.00% 76.99% -0.01%
==========================================
Files 121 121
Lines 7706 7708 +2
==========================================
+ Hits 5934 5935 +1
- Misses 1772 1773 +1
Continue to review full report at Codecov.
|
Nice example! Any idea why the minimum times show a much more pronounced speed-up? It could be because |
@maleadt great question. I added |
@maleadt I added another example that does a reduction over a large array. Surprisingly, the multiple GPU case is significantly slower (although the maximum time is about 1/3 the single GPU case). Perhaps there is a better way to partition the data/computation than I'm doing here? |
7a2b21f
to
55b8716
Compare
a1af3c1
to
46f6109
Compare
Nice examples ! |
As I work through how to speed-up some of functionality in the SciML codebases using multiple GPUs, I thought I'd add my small experiments as examples for other users of this package. Comments/feedback is welcome if the example(s) shown could be done better.