Integrating Bayesian Network Structure into Residual Flows and Variational Autoencoders

Abstract

Deep generative models have become more popular in recent years due to their scalability and representation capacity. Unlike probabilistic graphical models, they typically do not incorporate specific domain knowledge. As such, this work explores incorporating arbitrary dependency structures, as specified by Bayesian networks, into variational autoencoders (VAEs). This is achieved by developing a new type of graphical normalizing flow, which extends residual flows by encoding conditional independence through masking of the flow’s residual block weight matrices, and using these to extend both the prior and inference network of the VAE. We show that the proposed graphical VAE provides a more interpretable model that generalizes better in data-sparse settings, when practitioners know or can hypothesize about certain latent factors in their domain. Furthermore, we show that graphical residual flows provide not only density estimation and inference performance competitive with existing graphical flows, but also more stable and accurate inversion in practice as a byproduct of the flow’s Lipschitz bounds.

Cite

Text

Mouton and Kroon. "Integrating Bayesian Network Structure into Residual Flows and Variational Autoencoders." Transactions on Machine Learning Research, 2023.

Markdown

[Mouton and Kroon. "Integrating Bayesian Network Structure into Residual Flows and Variational Autoencoders." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/mouton2023tmlr-integrating/)

BibTeX

@article{mouton2023tmlr-integrating,
  title     = {{Integrating Bayesian Network Structure into Residual Flows and Variational Autoencoders}},
  author    = {Mouton, Jacobie and Kroon, Rodney Stephen},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/mouton2023tmlr-integrating/}
}