Keras/TF implementation of AdamW, SGDW, NadamW, Warm Restarts, and Learning Rate multipliers
-
Updated
Jan 6, 2022 - Python
Keras/TF implementation of AdamW, SGDW, NadamW, Warm Restarts, and Learning Rate multipliers
[Python] [arXiv/cs] Paper "An Overview of Gradient Descent Optimization Algorithms" by Sebastian Ruder
A comparison between implementations of different gradient-based optimization algorithms (Gradient Descent, Adam, Adamax, Nadam, Amsgrad). The comparison was made on some of the most common functions used for testing optimization algorithms.
Фреймворк глубоко обучения на Numpy, написанный с целью изучения того, как все работает под "капотом".
Add a description, image, and links to the nadam topic page so that developers can more easily learn about it.
To associate your repository with the nadam topic, visit your repo's landing page and select "manage topics."