Skip to content

This project outlines 4 experiments to explore the effects of several settings on the bias-variance tradeoff curve

Notifications You must be signed in to change notification settings

nplotko/double-descent-lr

Repository files navigation

Experiments with Overparameterized Models with Linear Regression

In this repository, I conduct 4 different experiments to view how the double descent curve changes. The double descent curve is a phenomenon dealing with overparameterized models. In certain settings of neural networks, the global optimum can be found in the overparameterized setting, so I try to experiment with this with Linear Regression. The data I work with is randomly generated (Normal).

Files:
4803_ProjectReport: Paper outlining the work I tried to replicate and the phenomenon in depth.

Experiment 1: This experiment varies the number of samples and holds the number of dimensions constant. I replicated the experiment of Preetum Nakkiran in "More Data Can Hurt in Linear Regression: Sample-wise Double Descent." Here, I tried two settings for the number of dimensions/parameters: 1000 and 500. This did not make a difference, as the shape of the curve is a function of the ratio of n (the number of samples) to p (the number of parameters).

Experiment 2: This experiment varies the number of dimensions (parameters) and holds the number of samples constant. I do this for 100, 500, and 1000 samples. Once, again the ratio of samples to dimensions is the important part here, so the curves don't look much different for each setting. I varied the signal-to-noise ratio (SNR) by changing the standard deviation of the data matrix, and this affects the shape of the curve.

Experiment 3: This experiment varies the number of dimensions (parameters) and holds the number of samples constant. I do this for 100, 500, and 1000 samples. Here, I misspecify the model by adding additional unnecessary samples and by leaving out significant features.

Experiment 4: This experiment varies the number of dimensions (parameters) and holds the number of samples constant. I do this for 100, 500, and 1000 samples. Here, I add a ridge regularization constant to view the effect of a regularization hyperparameter on the double descent curve.

About

This project outlines 4 experiments to explore the effects of several settings on the bias-variance tradeoff curve

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages