Learning to Drop Out: An Adversarial Approach to Training Sequence VAEs

Abstract

In principle, applying variational autoencoders (VAEs) to sequential data offers a method for controlled sequence generation, manipulation, and structured representation learning. However, training sequence VAEs is challenging: autoregressive decoders can often explain the data without utilizing the latent space, known as posterior collapse. To mitigate this, state-of-the-art models weaken' thepowerful decoder' by applying uniformly random dropout to the decoder input.We show theoretically that this removes pointwise mutual information provided by the decoder input, which is compensated for by utilizing the latent space. We then propose an adversarial training strategy to achieve information-based stochastic dropout. Compared to uniform dropout on standard text benchmark datasets, our targeted approach increases both sequence modeling performance and the information captured in the latent space.

Cite

Text

Miladinovic et al. "Learning to Drop Out: An Adversarial Approach to Training Sequence VAEs." Neural Information Processing Systems, 2022.

Markdown

[Miladinovic et al. "Learning to Drop Out: An Adversarial Approach to Training Sequence VAEs." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/miladinovic2022neurips-learning/)

BibTeX

@inproceedings{miladinovic2022neurips-learning,
  title     = {{Learning to Drop Out: An Adversarial Approach to Training Sequence VAEs}},
  author    = {Miladinovic, Djordje and Shridhar, Kumar and Jain, Kushal and Paulus, Max and Buhmann, Joachim M and Allen, Carl},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/miladinovic2022neurips-learning/}
}