Provable ICA with Unknown Gaussian Noise, with Implications for Gaussian Mixtures and Autoencoders

Abstract

We present a new algorithm for Independent Component Analysis (ICA) which has provable performance guarantees. In particular, suppose we are given samples of the form $y = Ax + \eta$ where $A$ is an unknown $n \times n$ matrix and $x$ is chosen uniformly at random from $\{+1, -1\}^n$, $\eta$ is an $n$-dimensional Gaussian random variable with unknown covariance $\Sigma$: We give an algorithm that provable recovers $A$ and $\Sigma$ up to an additive $\epsilon$ whose running time and sample complexity are polynomial in $n$ and $1 / \epsilon$. To accomplish this, we introduce a novel ``quasi-whitening'' step that may be useful in other contexts in which the covariance of Gaussian noise is not known in advance. We also give a general framework for finding all local optima of a function (given an oracle for approximately finding just one) and this is a crucial step in our algorithm, one that has been overlooked in previous attempts, and allows us to control the accumulation of error when we find the columns of $A$ one by one via local search.

Cite

Text

Arora et al. "Provable ICA with Unknown Gaussian Noise, with Implications for Gaussian Mixtures and Autoencoders." Neural Information Processing Systems, 2012.

Markdown

[Arora et al. "Provable ICA with Unknown Gaussian Noise, with Implications for Gaussian Mixtures and Autoencoders." Neural Information Processing Systems, 2012.](https://mlanthology.org/neurips/2012/arora2012neurips-provable/)

BibTeX

@inproceedings{arora2012neurips-provable,
  title     = {{Provable ICA with Unknown Gaussian Noise, with Implications for Gaussian Mixtures and Autoencoders}},
  author    = {Arora, Sanjeev and Ge, Rong and Moitra, Ankur and Sachdeva, Sushant},
  booktitle = {Neural Information Processing Systems},
  year      = {2012},
  pages     = {2375-2383},
  url       = {https://mlanthology.org/neurips/2012/arora2012neurips-provable/}
}