-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
It seems that the model has not learned anything,What should I do? #11
Comments
When I use linear_evaluation to evaluate my model,should I train on main_lincls.py first? |
I think you only need to change the dataloader and augmentation to CIFAR-10 and change the num_classes=10. And check if you can load the pretrained weights correctly. |
What is the right order of training? |
I'm confused about the order in which main.py and main_lincls.py are used |
You should, 1. use main.py to get the checkpoint; 2. use main_lincls.py to load the checkpoint as pretrained weights (not resume training). The results of CIFAR-10 in the paper are produced using ImageNet pretrained weights. I didn't try directly pretraining on CIFAR-10. Actually, 37% precision of linear evaluation shows the model's weight is not random. It seems the model learns some features, but not that good. The reason may lie in inappropriate hyper-parameters. Or maybe the CIFAR-10 dataset is too easy to learn, which makes the model's outputs became unchanged during the training process. |
Oh,I see. I use the main_lincls.py to get the checkpoint and than use it to evaluate. |
In "linear", we load the pretrained weights and fix the backbone, then only train a classifier. In "fine-tune" we load the pretrained weights as initialization and train the whole model normally. They are two different ways to measure the quality of pretrained weights. |
Thanks for your excellent work!
I change the dataloader to use JigClu in CIFAR-10,and train the model on it by 1000epoch.
But the prediction of my model is all the same.
It seem that model always cluster into the same cluster
The text was updated successfully, but these errors were encountered: