New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem migrating from DBN to lasagne NeuralNet: NaN for each epoch #35
Comments
I have put together this simple example fitting nolearns NeuralNet on MNIST, which also doesn't run on my machine (nan for losses + valid acc does not improve). Could you try to run it?
|
I think what you're missing is |
Ah, thanks, yes that was it! Apparently I looked into every other parameter besides output_nonlinearity... thanks again! |
Lasagne looks fantastic, thanks for integrating it into nolearn! However, I have trouble transitioning from nolearn's DBN to the new lasagne NeuralNet.
Here is what happens:
Done loading and transforming data, traindata size: 83.5334777832 MB
Distribution of classes in train data:
[[ 0.00000000e+00 5.82160000e+04]
[ 1.00000000e+00 5.12730000e+04]] 2
conf: momentum: 0.01 self.learn_rates: 0.01
fitting classifier... nolearn
InputLayer (None, 200) produces 200 outputs
DenseLayer (None, 50) produces 50 outputs
DenseLayer (None, 2) produces 2 outputs
I tried fiddling with different learning rates (1,0.1,0.01,... 0.0000001 even 0.0), momentum rates, different optimisers (sgd,nestrov, rmsprop ...every method that lasagne offers), input sizes, no. of hidden units, two and one hidden layer, all to no avail.
The mnist example from lasagne runs fine though.
Here is my DBN code, which also runs fine and produces models with >0.90% accuracy (on an audio gender detection task), on the same data:
I've translated that into:
Is there anything obvious that I've missed here? How can I debug this?
The text was updated successfully, but these errors were encountered: