Skip to content

added AMSgrad optimizer to Adam and SparseAdam #4034

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 6 commits into from
Dec 18, 2017
Merged

Conversation

kashif
Copy link
Contributor

@kashif kashif commented Dec 5, 2017

  • need to look at SparseAMSGrad version

@kashif
Copy link
Contributor Author

kashif commented Dec 12, 2017

ok ready for a review!

@kashif kashif changed the title added AMSgrad optimizer added AMSgrad optimizer to Adam and SparseAdam Dec 12, 2017
@soumith
Copy link
Member

soumith commented Dec 18, 2017

@pytorchbot test this please

@soumith soumith merged commit 68c0998 into pytorch:master Dec 18, 2017
@soumith
Copy link
Member

soumith commented Dec 18, 2017

thanks @kashif !

@kashif kashif deleted the AMSGrad branch December 18, 2017 18:41
Copy link
Contributor

@apaszke apaszke left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think there's a bug in SparseAdam now

if amsgrad:
torch.max(old_max_exp_avg_sq_values, exp_avg_sq_update_values, out=old_max_exp_avg_sq_values)
denom = old_max_exp_avg_sq_values.sqrt_().add_(group['eps'])
max_exp_avg_sq = make_sparse(old_max_exp_avg_sq_values)

This comment was marked as off-topic.

This comment was marked as off-topic.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants