Skip to content

code for Asymptotics of feature learning in two-layer networks after one gradient-step

Notifications You must be signed in to change notification settings

SPOC-group/OneStepGD_asymptotics

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 

Repository files navigation

OneStepGD_asymptotics

code for Asymptotics of feature learning in two-layer networks after one gradient-step

$\texttt{theory.ipynb}$ contains the numerical implementation of the theoretical characterization of Result 3.3, taking as inputs the normalized learning rate $\tilde{\eta}$, sample complexity $\alpha_0$ for the first gradient step, activations $\sigma,\sigma_\star$, and readout regularization $\lambda$.

$\texttt{Simulations.ipynb}$ implements the corresponding numerical experiments, namely one large gradient-descent step on the first layer weights, followed by ridge regression on the readout weights.

About

code for Asymptotics of feature learning in two-layer networks after one gradient-step

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages