[Discussion] What to improve on ConvNeXt? #1093
Replies: 2 comments
-
From my experience of training on Imagenet and architecture design there are several things that could be beneficial for speed.
btw i'm open for collaboration. i have ideas, code and 100+ experiments done on imagenet so far, but i'm short on GPUs recently so had to stop the research. feel free to email me at |
Beta Was this translation helpful? Give feedback.
-
Hi @bonlime , Let's keep the conversation here for now, I think it will be beneficial as there may already be people doing the same thing or people with better suggestions on how to move forward. I hope to not end up wasting GPU resources on the same idea. I will get back to this thread over the weekend when I am free. I am very interested to read the papers that you have mentioned. GPUs avaialbility should not be a problem. |
Beta Was this translation helpful? Give feedback.
-
Hi all,
In reference to the ConvNeXt paper, what do you think we can improve?
It will be great if we can continue to add more improvements that has already been proven by other cnn/transformers research papers to improve performance or reduce inference time of the model.
An obvious one will be to incorporate an auxiliary attention layer such as Convolutional Block Attention Module (CBAM). Another one that comes to mind is edit some blocks to be parallel convolutions with repeated fusion (like HRNet).
Hi @rwightman , I am interested to help out with training if there are any good ideas to improve the model. I can contribute the trained weights once it is done.
It will be great to consolidate more proven ideas together for CNN to properly benchmark a "modern" CNN to a vision transformer.
Beta Was this translation helpful? Give feedback.
All reactions