Parallel Multiscale Autoregressive Density Estimation

Abstract

PixelCNN achieves state-of-the-art results in density estimation for natural images. Although training is fast, inference is costly, requiring one network evaluation per pixel; O(N) for N pixels. This can be sped up by caching activations, but still involves generating each pixel sequentially. In this work, we propose a parallelized PixelCNN that allows more efficient inference by modeling certain pixel groups as conditionally independent. Our new PixelCNN model achieves competitive density estimation and orders of magnitude speedup – O(log N) sampling instead of O(N) – enabling the practical generation of 512x512 images. We evaluate the model on class-conditional image generation, text-to-image synthesis, and action-conditional video generation, showing that our model achieves the best results among non-pixel-autoregressive density models that allow efficient sampling.

Cite

Text

Reed et al. "Parallel Multiscale Autoregressive Density Estimation." International Conference on Machine Learning, 2017.

Markdown

[Reed et al. "Parallel Multiscale Autoregressive Density Estimation." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/reed2017icml-parallel/)

BibTeX

@inproceedings{reed2017icml-parallel,
  title     = {{Parallel Multiscale Autoregressive Density Estimation}},
  author    = {Reed, Scott and Oord, Aäron and Kalchbrenner, Nal and Colmenarejo, Sergio Gómez and Wang, Ziyu and Chen, Yutian and Belov, Dan and Freitas, Nando},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {2912-2921},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/reed2017icml-parallel/}
}