-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BN on test set #2
Comments
Totally true .. I guess the batch size of 100 gives "good enough" statistics for the problem so I forgot to add it in. Will try to update with a version that stores population statistics and properly uses those at test time. |
TF slim has population statistics recurrent batch norm you can check out I like your implementation style more though since it is elegant and in You might also want to play around with the random permutation mnist task On Tuesday, August 30, 2016, Olav Nymoen notifications@github.com wrote:
|
I've tried running with population statistics a bit now with really poor results on sequential mnist. Same results when using slim.batch_norm. The model seems to be dependent on the batch normalization. To test I tried using local batch statistics, but increasing the batch from 100 to 1000. That works better than full population statistics, but much worse than batch statistics for a batch of 100. The graphs in the paper looks very much like mine when using local batch statistics, however they explicitly mention using population statistics for their final results, so I'm not sure what's going on in my code. |
I've been trying the same recently, and share similar frustrations as you. But I think in the paper, the actual statistics are recorded separately at I also got really good results just using vanilla LSTM but initializing the On Saturday, September 3, 2016, Olav Nymoen notifications@github.com
|
@hardmaru Recently I also got worse result in test data set use pop mean and variance. You said you got good result just using vanilla lstm, could you please share you code and tell me what's going on.
|
Nice blog post!
One comment: When you evaluate the validation/test set, you should use the saved statistics from training. Looking at the code, I think you are calculating the moments as well during validation/test runs.
The text was updated successfully, but these errors were encountered: