-
Notifications
You must be signed in to change notification settings - Fork 6.8k
update the performance page of MXNet. #10761
Conversation
@mli do you want to review and merge it? |
@zheng-da did you set KM Affinity when test CPU performance? |
@pengzhao-intel yes, i did. Do you find anything unexpected? |
| 32 | 4883.77 | 854.4 | 1197.74 | 493.72 | 713.17 | 294.17 | | ||
| Batch | Alexnet | VGG | Inception-BN | Inception-v3 | Resnet 50 | Resnet 152 | | ||
|-------|---------|--------|--------------|--------------|-----------|------------| | ||
| 1 | 243.93 | 43.59 | 68.62 | 35.52 | 67.41 | 23.65 | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems like we're having a few regressions :(
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
First, @mli told me that M60 is supposed to be slower than M40.
VGG is slower because a different model was used. The original performance was measured a long time ago. Since then, the implementation of VGG has changed. The current version of VGG has many more layers. If you want to know more details, I think @TaoLv can tell you more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Right. benchmark_score.py
was changed last December (commit) and VGG test was updated from VGG-11 to VGG-16. Perf numbers in this PR are measured on VGG-16 and previous perf numbers of MKLML were measured on VGG-11.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actually, this change in VGG applies to all benchmark results. Not just MKLML vs. MKLDNN.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for updating the chart! It seems like we're having some regressions on the GPU version of MXNet. It would be great if somebody could follow up on those.
That's good. For the large BS=32, the performance of Alexnet/VGG/inception-xx have a sligt drop.
|
@pengzhao-intel I'm not sure why the performance for large batch sizes gets worse. It seems to me that your performance was measured after the PRs that improved the performance of MKLDNN a while ago (otherwise, Alexnet should have much worse performance). However, the performance I just measured on C5.18x matches the performance I saw when I wrote the blog. |
SW : the master branch of mxnet, commit id: 48749a5 |
The commit was two months ago. It's surprising that the performance was better for large batch sizes. I'll try it again tomorrow. |
@pengzhao-intel Here is the performance result on C5.18x for commit id: 48749a5. It's different from yours. Before running the benchmark, I set the thread affinity and the number of OMP threads as below. Do you see anything wrong?
|
The setting is same with ours. From your log: So, the Alexnet and Inception data are similar but VGG is different, right? @huangzhiyuan can provide more details. |
This reverts commit ebd8a6b.
Description
This updates the performance page of MXNet on CPU and GPU.
Checklist
Essentials
Please feel free to remove inapplicable items for your PR.
Changes
Comments