Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Weight Decay #14

Closed
cancan101 opened this issue Dec 31, 2014 · 4 comments · Fixed by #50
Closed

Weight Decay #14

cancan101 opened this issue Dec 31, 2014 · 4 comments · Fixed by #50

Comments

@cancan101
Copy link
Contributor

Add support for weight decay to nolearn.

It looks like lasagne has support although it might need some more work.
Also see cuda-convnet docs.

@cancan101
Copy link
Contributor Author

I think the loss function will need to have its signature updated to support this.

@dnouri
Copy link
Owner

dnouri commented Jan 5, 2015

Yes, I came across this myself recently; wanted to implement a cost function that depends on weights, so like an L2. Maybe it could be called objective, and the loss be deprecated. Maybe the arguments should be something like (layers, X_batch, y_batch), with layers being the output of self.get_all_layers()?

@cancan101
Copy link
Contributor Author

This is what pylearn2 does:

layer_costs = [ wrapped_layer_cost(layer, coeff)
   for layer, coeff in safe_izip(model.layers, self.coeffs) ]

so it looks like self.get_all_layers() is likely enough.

@dnouri
Copy link
Owner

dnouri commented Feb 8, 2015

We should try and use whatever Lasagne comes up with as part of fixing its regularization story. See Lasagne/Lasagne#14

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants