What Makes for Good Image Captions?

Abstract

This paper establishes a formal information-theoretic framework for image captioning, conceptualizing captions as compressed linguistic representations that selectively encode semantic units in images. Our framework posits that good image captions should balance three key aspects: informationally sufficient, minimally redundant, and readily comprehensible by humans. By formulating these aspects as quantitative measures with adjustable weights, our framework provides a flexible foundation for analyzing and optimizing image captioning systems across diverse task requirements. To demonstrate its applicability, we introduce the Pyramid of Captions (PoCa) method, which generates enriched captions by integrating local and global visual information. We present both theoretical proof that PoCa improves caption quality under certain assumptions, and empirical validation of its effectiveness across various image captioning models and datasets.

Cite

Text

Chen et al. "What Makes for Good Image Captions?." NeurIPS 2024 Workshops: Compression, 2024.

Markdown

[Chen et al. "What Makes for Good Image Captions?." NeurIPS 2024 Workshops: Compression, 2024.](https://mlanthology.org/neuripsw/2024/chen2024neuripsw-makes/)

BibTeX

@inproceedings{chen2024neuripsw-makes,
  title     = {{What Makes for Good Image Captions?}},
  author    = {Chen, Delong and Cahyawijaya, Samuel and Ishii, Etsuko and Chan, Ho Shu and Bang, Yejin and Fung, Pascale},
  booktitle = {NeurIPS 2024 Workshops: Compression},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/chen2024neuripsw-makes/}
}