Spatial Broadcast Decoder: A Simple Architecture for Disentangled Representations in VAEs

Abstract

We present a neural rendering architecture that helps variational autoencoders (VAEs) learn disentangled representations. Instead of the deconvolutional network typically used in the decoder of VAEs, we tile (broadcast) the latent vector across space, concatenate fixed X- and Y-“coordinate” channels, and apply a fully convolutional network with 1x1 stride. This provides an architectural prior for dissociating positional from non-positional features in the latent space, yet without providing any explicit supervision to this effect. We show that this architecture, which we term the Spatial Broadcast decoder, improves disentangling, reconstruction accuracy, and generalization to held-out regions in data space. We show the Spatial Broadcast Decoder is complementary to state-of-the-art (SOTA) disentangling techniques and when incorporated improves their performance.

Cite

Text

Watters et al. "Spatial Broadcast Decoder: A Simple Architecture for Disentangled Representations in VAEs." ICLR 2019 Workshops: LLD, 2019.

Markdown

[Watters et al. "Spatial Broadcast Decoder: A Simple Architecture for Disentangled Representations in VAEs." ICLR 2019 Workshops: LLD, 2019.](https://mlanthology.org/iclrw/2019/watters2019iclrw-spatial/)

BibTeX

@inproceedings{watters2019iclrw-spatial,
  title     = {{Spatial Broadcast Decoder: A Simple Architecture for Disentangled Representations in VAEs}},
  author    = {Watters, Nick and Matthey, Loic and Burgess, Chris P. and Lerchner, Alexander},
  booktitle = {ICLR 2019 Workshops: LLD},
  year      = {2019},
  url       = {https://mlanthology.org/iclrw/2019/watters2019iclrw-spatial/}
}