Random Consensus Robust PCA

Abstract

This paper presents r2pca , a random consensus method for robust principal component analysis. r2pca takes ransac ’s principle of using as little data as possible one step further. It iteratively selects small subsets of the data to identify pieces of the principal components, to then stitch them together. We show that if the principal components are in general position and the errors are suffi-ciently sparse, r2pca will exactly recover the principal components with probability 1, in lieu of assumptions on coherence or the distribution of the sparse errors, and even under adversarial settings. r2pca enjoys many advantages: it works well under noise, its computational complexity scales linearly in the ambient dimension, it is easily parallelizable, and due to its low sample complexity, it can be used in settings where data is so large it cannot even be stored in memory. We complement our theoretical findings with synthetic and real data experiments showing that r2pca outperforms state-of-the-art methods in a broad range of settings.

Cite

Text

Pimentel-Alarcón and Nowak. "Random Consensus Robust PCA." International Conference on Artificial Intelligence and Statistics, 2017. doi:10.1214/17-EJS1377SI

Markdown

[Pimentel-Alarcón and Nowak. "Random Consensus Robust PCA." International Conference on Artificial Intelligence and Statistics, 2017.](https://mlanthology.org/aistats/2017/pimentelalarcon2017aistats-random/) doi:10.1214/17-EJS1377SI

BibTeX

@inproceedings{pimentelalarcon2017aistats-random,
  title     = {{Random Consensus Robust PCA}},
  author    = {Pimentel-Alarcón, Daniel L. and Nowak, Robert D.},
  booktitle = {International Conference on Artificial Intelligence and Statistics},
  year      = {2017},
  pages     = {344-352},
  doi       = {10.1214/17-EJS1377SI},
  url       = {https://mlanthology.org/aistats/2017/pimentelalarcon2017aistats-random/}
}