Co-Attention CNNs for Unsupervised Object Co-Segmentation
Abstract
Object co-segmentation aims to segment the common objects in images. This paper presents a CNN-based method that is unsupervised and end-to-end trainable to better solve this task. Our method is unsupervised in the sense that it does not require any training data in the form of object masks but merely a set of images jointly covering objects of a specific class. Our method comprises two collaborative CNN modules, a feature extractor and a co-attention map generator. The former module extracts the features of the estimated objects and backgrounds, and is derived based on the proposed co-attention loss which minimizes inter-image object discrepancy while maximizing intra-image figure-ground separation. The latter module is learned to generated co-attention maps by which the estimated figure-ground segmentation can better fit the former module. Besides, the co-attention loss, the mask loss is developed to retain the whole objects and remove noises. Experiments show that our method achieves superior results, even outperforming the state-of-the-art, supervised methods.
Cite
Text
Hsu et al. "Co-Attention CNNs for Unsupervised Object Co-Segmentation." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/104Markdown
[Hsu et al. "Co-Attention CNNs for Unsupervised Object Co-Segmentation." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/hsu2018ijcai-co/) doi:10.24963/IJCAI.2018/104BibTeX
@inproceedings{hsu2018ijcai-co,
title = {{Co-Attention CNNs for Unsupervised Object Co-Segmentation}},
author = {Hsu, Kuang-Jui and Lin, Yen-Yu and Chuang, Yung-Yu},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {748-756},
doi = {10.24963/IJCAI.2018/104},
url = {https://mlanthology.org/ijcai/2018/hsu2018ijcai-co/}
}