Learning Diverse Image Colorization

Abstract

Colorization is an ambiguous problem, with multiple viable colorizations for a single grey-level image. However, previous methods only produce the single most probable colorization. Our goal is to model the diversity intrinsic to the problem of colorization and produce multiple colorizations that display long-scale spatial co-ordination. We learn a low dimensional embedding of color fields using a variational autoencoder (VAE). We construct loss terms for the VAE decoder that avoid blurry outputs and take into account the uneven distribution of pixel colors. Finally, we build a conditional model for the multi-modal distribution between grey-level image and the color field embeddings. Samples from this conditional model result in diverse colorization. We demonstrate that our method obtains better diverse colorizations than a standard conditional variational autoencoder (CVAE) model, as well as a recently proposed conditional generative adversarial network (cGAN).

Cite

Text

Deshpande et al. "Learning Diverse Image Colorization." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.307

Markdown

[Deshpande et al. "Learning Diverse Image Colorization." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/deshpande2017cvpr-learning/) doi:10.1109/CVPR.2017.307

BibTeX

@inproceedings{deshpande2017cvpr-learning,
  title     = {{Learning Diverse Image Colorization}},
  author    = {Deshpande, Aditya and Lu, Jiajun and Yeh, Mao-Chuang and Chong, Min Jin and Forsyth, David},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2017},
  doi       = {10.1109/CVPR.2017.307},
  url       = {https://mlanthology.org/cvpr/2017/deshpande2017cvpr-learning/}
}