Parallel Multiscale Autoregressive Density Estimation
Abstract
PixelCNN achieves state-of-the-art results in density estimation for natural images. Although training is fast, inference is costly, requiring one network evaluation per pixel; O(N) for N pixels. This can be sped up by caching activations, but still involves generating each pixel sequentially. In this work, we propose a parallelized PixelCNN that allows more efficient inference by modeling certain pixel groups as conditionally independent. Our new PixelCNN model achieves competitive density estimation and orders of magnitude speedup – O(log N) sampling instead of O(N) – enabling the practical generation of 512x512 images. We evaluate the model on class-conditional image generation, text-to-image synthesis, and action-conditional video generation, showing that our model achieves the best results among non-pixel-autoregressive density models that allow efficient sampling.
Cite
Text
Reed et al. "Parallel Multiscale Autoregressive Density Estimation." International Conference on Machine Learning, 2017.Markdown
[Reed et al. "Parallel Multiscale Autoregressive Density Estimation." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/reed2017icml-parallel/)BibTeX
@inproceedings{reed2017icml-parallel,
title = {{Parallel Multiscale Autoregressive Density Estimation}},
author = {Reed, Scott and Oord, Aäron and Kalchbrenner, Nal and Colmenarejo, Sergio Gómez and Wang, Ziyu and Chen, Yutian and Belov, Dan and Freitas, Nando},
booktitle = {International Conference on Machine Learning},
year = {2017},
pages = {2912-2921},
volume = {70},
url = {https://mlanthology.org/icml/2017/reed2017icml-parallel/}
}