On the Number of Variables to Use in Principal Component Regression

Abstract

We study least squares linear regression over $N$ uncorrelated Gaussian features that are selected in order of decreasing variance. When the number of selected features $p$ is at most the sample size $n$, the estimator under consideration coincides with the principal component regression estimator; when $p>n$, the estimator is the least $\ell_2$ norm solution over the selected features. We give an average-case analysis of the out-of-sample prediction error as $p,n,N \to \infty$ with $p/N \to \alpha$ and $n/N \to \beta$, for some constants $\alpha \in [0,1]$ and $\beta \in (0,1)$. In this average-case setting, the prediction error exhibits a ``double descent'' shape as a function of $p$. We also establish conditions under which the minimum risk is achieved in the interpolating ($p>n$) regime.

Cite

Text

Xu and Hsu. "On the Number of Variables to Use in Principal Component Regression." Neural Information Processing Systems, 2019.

Markdown

[Xu and Hsu. "On the Number of Variables to Use in Principal Component Regression." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/xu2019neurips-number/)

BibTeX

@inproceedings{xu2019neurips-number,
  title     = {{On the Number of Variables to Use in Principal Component Regression}},
  author    = {Xu, Ji and Hsu, Daniel J.},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {5094-5103},
  url       = {https://mlanthology.org/neurips/2019/xu2019neurips-number/}
}