A Multi-Task Learning Approach for Image Captioning
Abstract
In this paper, we propose a Multi-task Learning Approach for Image Captioning (MLAIC ), motivated by the fact that humans have no difficulty performing such task because they possess capabilities of multiple domains. Specifically, MLAIC consists of three key components: (i) A multi-object classification model that learns rich category-aware image representations using a CNN image encoder; (ii) A syntax generation model that learns better syntax-aware LSTM based decoder; (iii) An image captioning model that generates image descriptions in text, sharing its CNN encoder and LSTM decoder with the object classification task and the syntax generation task, respectively. In particular, the image captioning model can benefit from the additional object categorization and syntax knowledge. To verify the effectiveness of our approach, we conduct extensive experiments on MS-COCO dataset. The experimental results demonstrate that our model achieves impressive results compared to other strong competitors.
Cite
Text
Zhao et al. "A Multi-Task Learning Approach for Image Captioning." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/168Markdown
[Zhao et al. "A Multi-Task Learning Approach for Image Captioning." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/zhao2018ijcai-multi-a/) doi:10.24963/IJCAI.2018/168BibTeX
@inproceedings{zhao2018ijcai-multi-a,
title = {{A Multi-Task Learning Approach for Image Captioning}},
author = {Zhao, Wei and Wang, Benyou and Ye, Jianbo and Yang, Min and Zhao, Zhou and Luo, Ruotian and Qiao, Yu},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {1205-1211},
doi = {10.24963/IJCAI.2018/168},
url = {https://mlanthology.org/ijcai/2018/zhao2018ijcai-multi-a/}
}