-
Notifications
You must be signed in to change notification settings - Fork 90
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About CUB-200-2011's accuracy #2
Comments
I reproduce TransFG code follow the instruction with overleap on 4 Tesla GPU V-100 entirely. I just got 91.2%(Paper: 91.7%) Besides, I add AutoAug in training stage as paper mentioned, I got a lower, 90.8% accuracy. Is there any details I missed? |
I reproduce the code and got the accuracy 90.8% with non-overleap, thus your result 91.2% is reasonable according to the paper that overleap is better than non-overleap. As for the AutoAug, according to the code of CUB data utils, there is no AutoAug. By the way, have you got any good results on Stanfor Car, Stanfor Dogs? |
I have not try other datasets yet. There is no AutoAug in official code indeed, however, Auto Aug is mentioned in TransFG paper training details. By the way, have you ever reproduce ViT on CUB using this training code? |
I have not reproduce VIT on CUB using TransFG, but I think the accuracy will a bit lower than 90.8% and 91.2% |
Similarly, I can't reproduce 91.7% on CUB-200-2011. Only got 91.0% and 90.8% with/without overlapping. |
I reproduce ViT on CUB using the reference code:https://github.com/jeonsworld/ViT-pytorch, and got the accuracy 90.7%. |
Hi guys, sorry for the late reply. I'm very busy with the ongoing competitions and other projects so I don't have much time to maintain this repo at this time. I checked the commit log and found that the problem is that when I cleaned my code the scale for contrastive loss is wrong and the model needs to load the pre-trained layer norm weights instead of learning from scratch. I've fixed the bugs and tested that for me I can now get around 91.6%-91.7% accuracy on CUB. Also, since I did not carefully search for the best hyper-parameters and just try the four recommended settings as in the original ViT paper so maybe you can have a try and share here if you find something better. Thanks! |
Hi, @jokerpwn. Could you please share the setting you used to get 90.7% for ViT? I've tried the four recommended settings in the original paper and the best I can get is 90.3%. If so, I'll update the paper. Thanks! |
I guess it's because I used only one GPU, and the other settings are no different. |
I reproduce ViT on CUB using the reference https://github.com/rwightman/pytorch-image-models, and got an accuracy |
Hi, @jokerpwn ,what is your setting to get 90.7% for ViT? what are four recommended settings on CUB? Is there any details I missed? |
@TACJu After you fixed the norm layer pretrained and contrastive loss, I can still not reproduce TrasnFG 91.7%. Only get 91.2% / 91.0% with and without overlap on CUB. Could you please check the code again or provide the training log? If anyone have reproduced the correct results, please leave a message below. |
hi, Did you train the CUB-200-2011 dataset for 10000 epoch really? |
Thanks for your work and sharing your codes! However, when I reproduce your code on 4 Tesla GPU V-100 entirely following the instruction with non-overleap, I just got 90.8% accuracy. Could you analyze the problem about this?
The text was updated successfully, but these errors were encountered: