{ "id": "2011.11477", "version": "v1", "published": "2020-11-23T15:38:50.000Z", "updated": "2020-11-23T15:38:50.000Z", "title": "Dimensionality reduction, regularization, and generalization in overparameterized regressions", "authors": [ "Ningyuan", "Huang", "David W. Hogg", "Soledad Villar" ], "categories": [ "stat.ML", "cs.LG" ], "abstract": "Overparameterization in deep learning is powerful: Very large models fit the training data perfectly and yet generalize well. This realization brought back the study of linear models for regression, including ordinary least squares (OLS), which, like deep learning, shows a \"double descent\" behavior. This involves two features: (1) The risk (out-of-sample prediction error) can grow arbitrarily when the number of samples $n$ approaches the number of parameters $p$, and (2) the risk decreases with $p$ at $p>n$, sometimes achieving a lower value than the lowest risk at $p