Greater performance loss when training on padded MNIST #27
zhuokaizhao
started this conversation in
General
Replies: 1 comment 2 replies
-
Hi @zhuokaizhao thanks for your nice words! :) That seems a weird behaviour and to be honest I don't see how the padding could make a difference here. Best, |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, first of all thank you for all the hard work you've put on this research project and make it very simple to use within PyTorch.
Recently I've been testing some models' translation/rotation equivariance properties. And I notice some unexpected behavior when testing models built with e2cnn and its non-e2cnn variant (for example, the tutorial you released https://github.com/QUVA-Lab/e2cnn/blob/master/examples/model.ipynb that works on MNIST). Since I would want to translate the image during testing, I prepared the padded-MNIST by padding each 28x28 MNIST image to 64x64 with constant zero padding to the right and bottom (so the original MNIST image is at the top left of the padded image). I trained the two networks and expected the training accuracy to be similar from both implementations. However, the non-e2cnn network performs similar to when training with the non-padded original MNIST (99% training accuracy), while the e2cnn model performs much worse (only 70% training accuracy).
It seems that the padding is the only change here. But I don't quite understand the reason. May I know if you had any insight why would padding influence the e2cnn's performance so much? Thank you very much.
Beta Was this translation helpful? Give feedback.
All reactions