Join GitHub today
GitHub is home to over 20 million developers working together to host and review code, manage projects, and build software together.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
Already on GitHub? Sign in to your account
Get NN working a bit better on the tests #16
Conversation
michaelhush
and others
added some commits
Nov 24, 2016
charmasaur
merged commit df56ca1
into
michaelhush:master
Dec 9, 2016
1 check passed
continuous-integration/travis-ci/pr
The Travis CI build passed
Details
charmasaur
referenced this pull request
Dec 9, 2016
Merged
Revert "Get NN working a bit better on the tests" #17
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
charmasaur commentedDec 9, 2016
I think the biggest problem was that the scipy minimizer wasn't actually doing any minimizing, I think because the landscape looked too flat for it (so it was exiting after 0 iterations). The consequence of this is that the neural network would only ever suggest either random new parameters or the current best-known parameters. Switching to provide the actual gradient via a function seems to have fixed this problem, and has improved the performance a lot -- now the net will suggest better parameters if it thinks it can.
I also added a bit of regularisation and improved the Adam learning rate. I'm a bit sad that this made a big difference, since I'd hoped we wouldn't need to tune those parameters, but we can investigate that more later.
Finally, I added an extra visualisation that shows the estimated NN cost surface (when there are two parameters).