-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
x1.5 model consistently underperforming #524
Comments
This may be due to the fact that the 1.5x model is not initialized with imagenet pre-training nanodet/nanodet/model/backbone/shufflenetv2.py Lines 9 to 10 in 3c9607c
|
Hi, and thanks for your response. I don't know how I could have overlooked this, but evidently I did.
So I also trained the same models on VISEM for 300 epochs, and although in the end, the x1.5 model has a higher mAP, it's because the x1.0 model drops in performance:
My interpretation is that the shufflenet backbone struggles with the nature of the images, as they are microscopic recordings, which an ImageNet pre-training does not generalize well to. Now, after looking into the matter of the missing weights for the backbone, I found the following issues: |
I'm using NanoDet-Plus for research purposes but ran into a weird issue where the x1.5 model variants consistently underperform compared to the x1.0 models. This happens on VISEM, Argoverse-HD, and even COCO2017.
For COCO2017 I used the stock config provided in the git repo (x1.0 and x1.5), but the COCO mAP already separates the two after the first 10 epochs, with the bigger variant consistently scoring lower. This holds true for all three datasets from my observations:
I get the following mAP metrics:
*The models are still training as of now, but the separation in mAP mentioned above is already distinctly noticeable in the logs. I will let the runs continue until 300 epochs have been reached, as the stock config dictates.
Here are the configs I used for VISEM and Argoverse-HD:
VISEM x1.0
VISEM x1.5
Argoverse-HD x1.0
Argoverse-HD x1.5
Is this erroneous behavior? Is there something wrong with my setup?
The text was updated successfully, but these errors were encountered: