Deep Descriptor Transforming for Image Co-Localization

Abstract

Reusable model design becomes desirable with the rapid expansion of machine learning applications. In this paper, we focus on the reusability of pre-trained deep convolutional models. Specifically, different from treating pre-trained models as feature extractors, we reveal more treasures beneath convolutional layers, i.e., the convolutional activations could act as a detector for the common object in the image co-localization problem. We propose a simple but effective method, named Deep Descriptor Transforming (DDT), for evaluating the correlations of descriptors and then obtaining the category-consistent regions, which can accurately locate the common object in a set of images. Empirical studies validate the effectiveness of the proposed DDT method. On benchmark image co-localization datasets, DDT consistently outperforms existing state-of-the-art methods by a large margin. Moreover, DDT also demonstrates good generalization ability for unseen categories and robustness for dealing with noisy data.

Cite

Text

Wei et al. "Deep Descriptor Transforming for Image Co-Localization." International Joint Conference on Artificial Intelligence, 2017. doi:10.24963/IJCAI.2017/425

Markdown

[Wei et al. "Deep Descriptor Transforming for Image Co-Localization." International Joint Conference on Artificial Intelligence, 2017.](https://mlanthology.org/ijcai/2017/wei2017ijcai-deep/) doi:10.24963/IJCAI.2017/425

BibTeX

@inproceedings{wei2017ijcai-deep,
  title     = {{Deep Descriptor Transforming for Image Co-Localization}},
  author    = {Wei, Xiu-Shen and Zhang, Chen-Lin and Li, Yao and Xie, Chen-Wei and Wu, Jianxin and Shen, Chunhua and Zhou, Zhi-Hua},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {3048-3054},
  doi       = {10.24963/IJCAI.2017/425},
  url       = {https://mlanthology.org/ijcai/2017/wei2017ijcai-deep/}
}