CT2: Colorization Transformer via Color Tokens

Abstract

Automatic image colorization is an ill-posed problem with multi-modal uncertainty, and there remains two main challenges with previous methods: incorrect semantic colors and under-saturation. In this paper, we propose an end-to-end transformer-based model to overcome these challenges. Benefited from the long-range context extraction of transformer and our holistic architecture, our method could colorize images with more diverse colors. Besides, we introduce color tokens into our approach and treat the colorization task as a classification problem, which increases the saturation of results. We also propose a series of modules to make image features interact with color tokens, and restrict the range of possible color candidates, which makes our results visually pleasing and reasonable. In addition, our method does not require any additional external priors, which ensures its well generalization capability. Extensive experiments and user studies demonstrate that our method achieves superior performance than previous works.

Cite

Text

Weng et al. "CT2: Colorization Transformer via Color Tokens." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-20071-7_1

Markdown

[Weng et al. "CT2: Colorization Transformer via Color Tokens." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/weng2022eccv-ct2/) doi:10.1007/978-3-031-20071-7_1

BibTeX

@inproceedings{weng2022eccv-ct2,
  title     = {{CT2: Colorization Transformer via Color Tokens}},
  author    = {Weng, Shuchen and Sun, Jimeng and Li, Yu and Li, Si and Shi, Boxin},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2022},
  doi       = {10.1007/978-3-031-20071-7_1},
  url       = {https://mlanthology.org/eccv/2022/weng2022eccv-ct2/}
}