StrWAEs to Invariant Representations

Abstract

Autoencoders have become an indispensable tool for generative modeling and representation learning in high dimensions. Imposing structural constraints such as conditional independence in order to capture invariance of latent variables to nuisance information has been attempted through adding ad hoc penalties to the loss function mostly in the variational autoencoder (VAE) context, often based on heuristics. This paper demonstrates that Wasserstein autoencoders (WAEs) are highly flexible in embracing such structural constraints. Well-known extensions of VAEs for this purpose are gracefully handled within the framework of WAEs. In particular, given a conditional independence structure of the generative model (decoder), corresponding encoder structure and penalties are derived from the functional constraints that define the WAE. These structural uses of WAEs, termed StrWAEs (“stairways”), open up a principled way of penalizing autoencoders to impose structural constraints. Utilizing these advantages, we present handful of results on semi-supervised classification, conditional generation, and invariant representation tasks.

Cite

Text

Lee et al. "StrWAEs to Invariant Representations." International Conference on Machine Learning, 2024.

Markdown

[Lee et al. "StrWAEs to Invariant Representations." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/lee2024icml-strwaes/)

BibTeX

@inproceedings{lee2024icml-strwaes,
  title     = {{StrWAEs to Invariant Representations}},
  author    = {Lee, Hyunjong and Seong, Yedarm and Lee, Sungdong and Won, Joong-Ho},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {26902-26928},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/lee2024icml-strwaes/}
}