Structure by Architecture: Structured Representations Without Regularization

Abstract

We study the problem of self-supervised structured representation learning using autoencoders for downstream tasks such as generative modeling. Unlike most methods which rely on matching an arbitrary, relatively unstructured, prior distribution for sampling, we propose a sampling technique that relies solely on the independence of latent variables, thereby avoiding the trade-off between reconstruction quality and generative performance typically observed in VAEs. We design a novel autoencoder architecture capable of learning a structured representation without the need for aggressive regularization. Our structural decoders learn a hierarchy of latent variables, thereby ordering the information without any additional regularization or supervision. We demonstrate how these models learn a representation that improves results in a variety of downstream tasks including generation, disentanglement, and extrapolation using several challenging and natural image datasets.

Cite

Text

Leeb et al. "Structure by Architecture: Structured Representations Without Regularization." International Conference on Learning Representations, 2023.

Markdown

[Leeb et al. "Structure by Architecture: Structured Representations Without Regularization." International Conference on Learning Representations, 2023.](https://mlanthology.org/iclr/2023/leeb2023iclr-structure/)

BibTeX

@inproceedings{leeb2023iclr-structure,
  title     = {{Structure by Architecture: Structured Representations Without Regularization}},
  author    = {Leeb, Felix and Lanzillotta, Giulia and Annadani, Yashas and Besserve, Michel and Bauer, Stefan and Schölkopf, Bernhard},
  booktitle = {International Conference on Learning Representations},
  year      = {2023},
  url       = {https://mlanthology.org/iclr/2023/leeb2023iclr-structure/}
}