EM Algorithms for PCA and SPCA

Abstract

I present an expectation-maximization (EM) algorithm for principal component analysis (PCA). The algorithm allows a few eigenvectors and eigenvalues to be extracted from large collections of high dimensional data. It is computationally very efficient in space and time. It also natu(cid:173) rally accommodates missing infonnation. I also introduce a new variant of PC A called sensible principal component analysis (SPCA) which de(cid:173) fines a proper density model in the data space. Learning for SPCA is also done with an EM algorithm. I report results on synthetic and real data showing that these EM algorithms correctly and efficiently find the lead(cid:173) ing eigenvectors of the covariance of datasets in a few iterations using up to hundreds of thousands of datapoints in thousands of dimensions.

Cite

Text

Roweis. "EM Algorithms for PCA and SPCA." Neural Information Processing Systems, 1997.

Markdown

[Roweis. "EM Algorithms for PCA and SPCA." Neural Information Processing Systems, 1997.](https://mlanthology.org/neurips/1997/roweis1997neurips-em/)

BibTeX

@inproceedings{roweis1997neurips-em,
  title     = {{EM Algorithms for PCA and SPCA}},
  author    = {Roweis, Sam T.},
  booktitle = {Neural Information Processing Systems},
  year      = {1997},
  pages     = {626-632},
  url       = {https://mlanthology.org/neurips/1997/roweis1997neurips-em/}
}