Hierarchical VAE with a Diffusion-Based VampPrior

Abstract

Deep hierarchical variational autoencoders (VAEs) are powerful latent variable generative models. In this paper, we introduce Hierarchical VAE with Diffusion-based Variational Mixture of the Posterior Prior (VampPrior). We apply amortization to scale the VampPrior to models with many stochastic layers. The proposed approach allows us to achieve better performance compared to the original VampPrior work and other deep hierarchical VAEs, while using fewer parameters. We empirically validate our method on standard benchmark datasets (MNIST, OMNIGLOT, CIFAR10) and demonstrate improved training stability and latent space utilization.

Cite

Text

Kuzina and Tomczak. "Hierarchical VAE with a Diffusion-Based VampPrior." Transactions on Machine Learning Research, 2024.

Markdown

[Kuzina and Tomczak. "Hierarchical VAE with a Diffusion-Based VampPrior." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/kuzina2024tmlr-hierarchical/)

BibTeX

@article{kuzina2024tmlr-hierarchical,
  title     = {{Hierarchical VAE with a Diffusion-Based VampPrior}},
  author    = {Kuzina, Anna and Tomczak, Jakub M.},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/kuzina2024tmlr-hierarchical/}
}