Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add adahessian #169

Closed
bratao opened this issue Aug 10, 2020 · 6 comments
Closed

Add adahessian #169

bratao opened this issue Aug 10, 2020 · 6 comments

Comments

@bratao
Copy link

bratao commented Aug 10, 2020

Hello,

I´m a big fan of this project. Recently a new optimized has been proposed, that promises SOTA for many tasks.

https://github.com/amirgholami/adahessian

The possibility of using it here is very good!

@jettify
Copy link
Owner

jettify commented Aug 10, 2020

Thanks! I will take a look.

@jettify
Copy link
Owner

jettify commented Aug 16, 2020

Since authors licensed code under unfriendly GPL V3, I can not reuse that code here and unfortunitly have to reproduce new implementation from paper.

Commercial usage of of authors implementation is limited, since forces users to license potentially commercial code under GPL also.

@bratao
Copy link
Author

bratao commented Aug 23, 2020

@jettify this is a re-implementation from scratch of the paper, with a MIT license
https://github.com/davda54/ada-hessian

@amirgholami
Copy link
Contributor

amirgholami commented Dec 13, 2020

Hi @jettify,

First, thanks so much for creating this repository, it is a great resource. I am one of the Adahessian authors, and we have just changed the license to the more friendly MIT License. I also just opened a new pull request that adds Adahessian optimizer to your repository:

#239 (comment)

I got the following results when I ran the optimizer on the Rosenbrock and Rastrigin test:

  • Rosenbrock: best loss: 0.002011366948950946, best lr= 32.26606775899826
  • Rastrigin: best loss: 2.3379755056396334e-05, best lr= 23.15826699019325 (best loss value among the current optimizers)
    P.S: Note that by design second order optimizers need larger learning rates

I have also attached the visualization for each test:

rosenbrock_Adahessian
rastrigin_Adahessian

P.S: Based on the results I got, it seems Adahessian gets the best loss value among the optimizers

P.S2: @bratao Thanks so much for adding the Adahessian optimizer and for promoting it.

@jettify
Copy link
Owner

jettify commented Dec 30, 2020

Just merged PR with optimizer, all thanks to @amirgholami !

@jettify jettify closed this as completed Dec 30, 2020
@jettify
Copy link
Owner

jettify commented Jan 1, 2021

PyPI is also updated https://pypi.org/project/torch-optimizer/0.1.0/

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants