-
Notifications
You must be signed in to change notification settings - Fork 74.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
hessian free optimizer needed #2682
Comments
PRs welcome! |
I would like to work on this. |
@ajaybhat Let us know if you have questions or issues! |
@girving Could you let me know which classes to look for as a reference to implement optimizers? I'm having a bit of trouble finding them. |
@ajaybhat Search for classes which inherit from |
Thanks! I'll let you know if i've any more questions. |
I was also try to implement HessianFree (HF) optimization on another framework, however i did not success. I am not sure if you can use this git (https://github.com/drasmuss/hessianfree) as reference or ask for help. The implementation is already on python and cuda, but he did not implemented convolutional. If you have a code to test let me know so i can compare with my others results. |
Is there anyone who is still trying to implement HF optimization on the tensorflow framework? |
I was attempting to implement HF optimisation and Saddle Free Newton. These algorithms are doable in a FFN framework. Unfortunately for RNN's as discussed in #5985 it is currently not possible to calculate second order derivatives from DynamicRNN's due to the use of a while loop. The current workaround would be to use StaticRNN. However, the while loop second derivative issue seems to be the major issue when attempting to implement a general Hessian Free optimisation algorithm or any other general second order method which requires Hessian Vector Products. |
Automatically closing due to lack of recent activity. Since this issue is old at this point, please reopen the issue if it still occurs when tried with the latest version of Tensorflow. Thank you. |
It's been more than a year since the Issue was closed, TF 1.3 was recently deployed and I still think that a Hessian Free Optimizer implementation for TF would be great. Consider reopening? |
Pardon me to promote my second-order optimization methods here. If interested, please check my tensorflow package at https://github.com/lixilinx/psgd_tf As for HF optimization, its damping factor and step size in line search are obtained by trial-and-error, and this could cause further troubles for its tensorflow implementation. A second-order method without line search is preferred, and methods in the above link are such examples. |
Here is an implementation of the Saddle Free method: |
amazing |
Hessian free optimizers are successfully applied on neural networks (especially RNNs). Currently need one in TF!
The text was updated successfully, but these errors were encountered: