An Information Geometry of Statistical Manifold Learning

Abstract

Manifold learning seeks low-dimensional representations of high-dimensional data. The main tactics have been exploring the geometry in an input data space and an output embedding space. We develop a manifold learning theory in a hypothesis space consisting of models. A model means a specific instance of a collection of points, e.g., the input data collectively or the output embedding collectively. The semi-Riemannian metric of this hypothesis space is uniquely derived in closed form based on the information geometry of probability distributions. There, manifold learning is interpreted as a trajectory of intermediate models. The volume of a continuous region reveals an amount of information. It can be measured to define model complexity and embedding quality. This provides deep unified perspectives of manifold learning theory.

Cite

Text

Sun and Marchand-Maillet. "An Information Geometry of Statistical Manifold Learning." International Conference on Machine Learning, 2014.

Markdown

[Sun and Marchand-Maillet. "An Information Geometry of Statistical Manifold Learning." International Conference on Machine Learning, 2014.](https://mlanthology.org/icml/2014/sun2014icml-information/)

BibTeX

@inproceedings{sun2014icml-information,
  title     = {{An Information Geometry of Statistical Manifold Learning}},
  author    = {Sun, Ke and Marchand-Maillet, Stéphane},
  booktitle = {International Conference on Machine Learning},
  year      = {2014},
  pages     = {1-9},
  volume    = {32},
  url       = {https://mlanthology.org/icml/2014/sun2014icml-information/}
}