Automatic Controllable Colorization via Imagination

Abstract

We propose a framework for automatic colorization that allows for iterative editing and modifications. The core of our framework lies in an imagination module: by understanding the content within a grayscale image we utilize a pre-trained image generation model to generate multiple images that contain the same content. These images serve as references for coloring mimicking the process of human experts. As the synthesized images can be imperfect or different from the original grayscale image we propose a Reference Refinement Module to select the optimal reference composition. Unlike most previous end-to-end automatic colorization algorithms our framework allows for iterative and localized modifications of the colorization results because we explicitly model the coloring samples. Extensive experiments demonstrate the superiority of our framework over existing automatic colorization algorithms in editability and flexibility. Project page: https://xy-cong.github.io/imagine-colorization/.

Cite

Text

Cong et al. "Automatic Controllable Colorization via Imagination." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.00252

Markdown

[Cong et al. "Automatic Controllable Colorization via Imagination." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/cong2024cvpr-automatic/) doi:10.1109/CVPR52733.2024.00252

BibTeX

@inproceedings{cong2024cvpr-automatic,
  title     = {{Automatic Controllable Colorization via Imagination}},
  author    = {Cong, Xiaoyan and Wu, Yue and Chen, Qifeng and Lei, Chenyang},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {2609-2619},
  doi       = {10.1109/CVPR52733.2024.00252},
  url       = {https://mlanthology.org/cvpr/2024/cong2024cvpr-automatic/}
}