Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning

Abstract

We study structure-preserving low-dimensional representation of a data manifold embedded in a high-dimensional observation space based on variational auto-encoders (VAEs). We approach this by decomposing the data manifold $\mathcal{M}$ as $\mathcal{M} = \mathcal{M} / G \times G$, where $G$ and $\mathcal{M} / G$ are a group of symmetry transformations and a quotient space of $\mathcal{M}$ up to $G$, respectively. From this perspective, we define the structure-preserving representation of such a manifold as a latent space $\mathcal{Z}$ which is isometrically isomorphic (i.e., distance-preserving) to the quotient space $\mathcal{M} / G$ rather $\mathcal{M}$ (i.e., symmetry-preserving). To this end, we propose a novel auto-encoding framework, named isometric quotient VAEs (IQVAEs), that can extract the quotient space from observations and learn the Riemannian isometry of the extracted quotient in an unsupervised manner. Empirical proof-of-concept experiments reveal that the proposed method can find a meaningful representation of the learned data and outperform other competitors for downstream tasks.

Cite

Text

Huh et al. "Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning." Neural Information Processing Systems, 2023.

Markdown

[Huh et al. "Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/huh2023neurips-isometric/)

BibTeX

@inproceedings{huh2023neurips-isometric,
  title     = {{Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning}},
  author    = {Huh, In and Jeong, Changwook and Choe, Jae Myung and Kim, Younggu and Kim, Daesin},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/huh2023neurips-isometric/}
}