Adaptive Metric Dimensionality Reduction

Abstract

We study data-adaptive dimensionality reduction in the context of supervised learning in general metric spaces. Our main statistical contribution is a generalization bound for Lipschitz functions in metric spaces that are doubling, or nearly doubling, which yields a new theoretical explanation for empirically reported improvements gained by preprocessing Euclidean data by PCA (Principal Components Analysis) prior to constructing a linear classifier. On the algorithmic front, we describe an analogue of PCA for metric spaces, namely an efficient procedure that approximates the data’s intrinsic dimension, which is often much lower than the ambient dimension. Our approach thus leverages the dual benefits of low dimensionality: (1) more efficient algorithms, e.g., for proximity search, and (2) more optimistic generalization bounds.

Cite

Text

Gottlieb et al. "Adaptive Metric Dimensionality Reduction." International Conference on Algorithmic Learning Theory, 2013. doi:10.1007/978-3-642-40935-6_20

Markdown

[Gottlieb et al. "Adaptive Metric Dimensionality Reduction." International Conference on Algorithmic Learning Theory, 2013.](https://mlanthology.org/alt/2013/gottlieb2013alt-adaptive/) doi:10.1007/978-3-642-40935-6_20

BibTeX

@inproceedings{gottlieb2013alt-adaptive,
  title     = {{Adaptive Metric Dimensionality Reduction}},
  author    = {Gottlieb, Lee-Ad and Kontorovich, Aryeh and Krauthgamer, Robert},
  booktitle = {International Conference on Algorithmic Learning Theory},
  year      = {2013},
  pages     = {279-293},
  doi       = {10.1007/978-3-642-40935-6_20},
  url       = {https://mlanthology.org/alt/2013/gottlieb2013alt-adaptive/}
}