Skip to content

ShiyunXu/NTK_Optimizer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 

Repository files navigation

NTK_Optimizer

This repo corresponds to my following paper: On the Optimization Algorithms of Over-parameterized Deep Neural Networks: A View from the Dynamical Systems

Abstract:

When equipped with efficient optimization algorithms, the over-parameterized neural networks have demonstrated high level of performance even though the loss function is non-convex and non-smooth. While many works have been focusing on understanding the loss dynamics by training neural networks with the gradient descent (GD), in this work, we consider a broad class of optimization algorithms that are commonly used in practice. For example, we show from a dynamical system perspective that the Heavy Ball (HB) method can converge to global minimum on mean squared error (MSE) at a linear rate (similar to GD); however, the Nesterov accelerated gradient descent (NAG) only converges to global minimum sublinearly.

Our results rely on the connection between neural tangent kernel (NTK) and finite over-parameterized neural networks with ReLU activation, which leads to analyzing the limiting ordinary differential equations (ODE) for optimization algorithms. We show that, optimizing the non-convex loss over the weights corresponds to optimizing some strongly convex loss over the prediction error. As a consequence, we can leverage the classical convex optimization theory to understand the convergence behavior of neural networks. We believe our approach can also be extended to other loss functions and network architectures.

Citation

Please cite our paper if you use NTK_Optimizier in your papers, as follows:

@inproceedings{bu2021dynamical,
  title={A dynamical view on optimization algorithms of overparameterized neural networks},
  author={Bu, Zhiqi and Xu, Shiyun and Chen, Kan},
  booktitle={International Conference on Artificial Intelligence and Statistics},
  pages={3187--3195},
  year={2021},
  organization={PMLR}
}

About

On the Optimization Algorithms of Over-parameterized Deep Neural Networks: A View from the Dynamical Systems

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published