-
Notifications
You must be signed in to change notification settings - Fork 80
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CIFAR-10 Reproduction #13
Comments
Hi @HieuPhan33 , Are you using an equivalent batch size of 1024? Could you try a smaller batch size like the default one (I usually use ~100, and found this to be important)? In addition, when I reproduced the result, I also sometimes (but generally rarely) get <93%, which is part of the fluctuation. If you still encounter the issue, you can also reach me via email and I can send you a sample training log for you to compare... I believe you should expect ~92% after 100 epochs already. |
Hi @jerrybai1995, thanks for quick response. I will reduce the batch size and keep you updated. |
Hi, I achieved 92.30% when using a batch size of 128. |
Hmmm, 92.3% still sounds too low to me for the given default parameters (my logs are usually in the range 92.6% - 93.4%). Could you try increasing f_thres (e.g., 9) and b_thres (e.g., 8 or 9) in the yaml file and using the default batch size? I also think that increasing the momentum (e.g., to 0.99) would improve the performance but I believe you should be able to reproduce the ~93% level performance even without tuning these things. I'll look into this but in case you might find it useful, feel free to contact me (shaojieb@cs.cmu.edu) and I'll send you some training logs. |
Hi @HieuPhan33 , I was able to produce 93.04% and 92.78% on two (slightly different and) independent runs, basically with the modifications/settings mentioned above. E.g., I got 93.04% from the following yaml:
Hope this helps! |
Thanks Shaojie, really appreciate your help! |
Hi Shaojie,
I could not reproduce the result for MDEQ on CIFAR-10 image classification.
I only obtained 91.56% using MDEQ_large.
I'm using the same parameters in
cls_mdeq_LARGE_reg.yaml
, except the batch size and the number of GPUs.Batch size per gpu are 512 with 2 GPUs.
I'm using 2 RTX 3090 graphic cards.
Hope that you can give me some advice.
Thanks Shaojie.
The text was updated successfully, but these errors were encountered: