-
Notifications
You must be signed in to change notification settings - Fork 20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
there is a bit of glitch about batch_size #2
Comments
I just replicated your problem, it is weird that the batch size affects performance so significantly, I suspect that it is due to #1. Yet have no idea why exactly this happens. |
emm…I see. I haven't notice the problem mentioned in #1 . Look forward to your fix. |
I just pushed an update in the dev branch, you can clone it and see what happens. I have run a few experiments but the issue still exists, and it seems irrelevant to the batch size. Let me summarize the problem (experiments below are all
So it seems that the loss surface has more than 2 local minimum and it is tricky to optimize to the global minimum. I have no idea why this happens, maybe I need to recheck every line of code to see if there is any wrong implementation. |
I just found that the problem may be caused by the wrong implementation of |
When I set the batch_size == 128, its performance downgraded significantly,namely its accuracy is about 15%.I wonder if there is any problem with the network implementation.
The text was updated successfully, but these errors were encountered: