New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SVHN - final accuracy #4
Comments
Hi, and thanks for letting me know. You should see around 94-95% accuracy against the validation set using that runner. I will try to reproduce your result. I don't know what's causing the discrepancy, but possibly I have made a mistake when I added additional experiments to the code (the consistency_trust, num_logits and logit_distance_cost hyperparams were not there when I ran the primary SVHN experiments that the results are based on) or when I cleaned up the code before publishing it. In the meanwhile, if you'd like to solve this yourself, you can try the version at commit In addition, experiments/svhn_final_eval.py should have the exact hyperparams that the experiments in the paper used. To make a single run, you can replace the parameters function with something like this:
But it is likely that the conceived bug also affects that runner. |
Hi. I looked into it and believe it's a hyperparameter issue. The runner uses However, I wasn't able to replicate the 90% result even with the wrong hyperparameter. I got consistently about 93% - 94% results. Maybe you hit a very unfortunate initialization there, or maybe there is some other issue. Somewhat tangentially, I am going to change the train_svhn.py hyperparameters so that it converges quickly but may not get quite optimal results. The svhn_final_eval.py will continue to be contain the hyperparameters in the paper that are close to optimal. |
Two other things to keep in mind:
|
I changed the hyperparams of I am closing this issue as I wasn't able to reproduce the 90% results, and because the runner is now clearly different from the one used in the paper. Try experiments/svhn_final_eval.py to reproduce the paper results. Please reopen if your results remain bad. |
Hi, I ran your tensorflow code (file train_svhn.py) and the final accuracy was only around 90%. I did not change anything in the code. I ran it as is ! Do you have any suggestions why I do not get the expected 96% ? By the way, I ran it on one GPU.
The text was updated successfully, but these errors were encountered: