New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Different learning rate per layer #10
Comments
Yes, you can specify different learning rate for each parameter, by scaling its gradient. You can overwrite the def get_gradient_processor(self):
return [ScaleGradient([('conv.*/W', 0.5), ('fc.*', 2)])] ScaleGradient takes a list of // UPDATE: this usage was later deprecated. See comment below. |
Exactly what I needed. Thank you! |
I couldn't find any function as |
This usage was deprecated one year ago. Now gradient processors can be added onto optimizer with http://tensorpack.readthedocs.io/en/latest/modules/tfutils.html#tensorpack.tfutils.optimizer.apply_grad_processors . |
For tensorflow you can use multiple optimizers to achieve different learning rates per layer.
Is there support for this in tensorpack or do I have to write a custom trainer?
The text was updated successfully, but these errors were encountered: