Linear Regression with Random Projections

Abstract

We investigate a method for regression that makes use of a randomly generated subspace GP⊂F (of finite dimension P) of a given large (possibly infinite) dimensional function space F, for example, L2([0,1]d;ℜ). GP is defined as the span of P random features that are linear combinations of a basis functions of F weighted by random Gaussian i.i.d. coefficients. We show practical motivation for the use of this approach, detail the link that this random projections method share with RKHS and Gaussian objects theory and prove, both in deterministic and random design, approximation error bounds when searching for the best regression function in GP rather than in F, and derive excess risk bounds for a specific regression algorithm (least squares regression in GP). This paper stresses the motivation to study such methods, thus the analysis developed is kept simple for explanations purpose and leaves room for future developments.

Cite

Text

Maillard and Munos. "Linear Regression with Random Projections." Journal of Machine Learning Research, 2012.

Markdown

[Maillard and Munos. "Linear Regression with Random Projections." Journal of Machine Learning Research, 2012.](https://mlanthology.org/jmlr/2012/maillard2012jmlr-linear/)

BibTeX

@article{maillard2012jmlr-linear,
  title     = {{Linear Regression with Random Projections}},
  author    = {Maillard, Odalric-Ambrym and Munos, Rémi},
  journal   = {Journal of Machine Learning Research},
  year      = {2012},
  pages     = {2735-2772},
  volume    = {13},
  url       = {https://mlanthology.org/jmlr/2012/maillard2012jmlr-linear/}
}