One-Shot Object Detection with Co-Attention and Co-Excitation

Abstract

This paper aims to tackle the challenging problem of one-shot object detection. Given a query image patch whose class label is not included in the training data, the goal of the task is to detect all instances of the same class in a target image. To this end, we develop a novel {\em co-attention and co-excitation} (CoAE) framework that makes contributions in three key technical aspects. First, we propose to use the non-local operation to explore the co-attention embodied in each query-target pair and yield region proposals accounting for the one-shot situation. Second, we formulate a squeeze-and-co-excitation scheme that can adaptively emphasize correlated feature channels to help uncover relevant proposals and eventually the target objects. Third, we design a margin-based ranking loss for implicitly learning a metric to predict the similarity of a region proposal to the underlying query, no matter its class label is seen or unseen in training. The resulting model is therefore a two-stage detector that yields a strong baseline on both VOC and MS-COCO under one-shot setting of detecting objects from both seen and never-seen classes.

Cite

Text

Hsieh et al. "One-Shot Object Detection with Co-Attention and Co-Excitation." Neural Information Processing Systems, 2019.

Markdown

[Hsieh et al. "One-Shot Object Detection with Co-Attention and Co-Excitation." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/hsieh2019neurips-oneshot/)

BibTeX

@inproceedings{hsieh2019neurips-oneshot,
  title     = {{One-Shot Object Detection with Co-Attention and Co-Excitation}},
  author    = {Hsieh, Ting-I and Lo, Yi-Chen and Chen, Hwann-Tzong and Liu, Tyng-Luh},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {2725-2734},
  url       = {https://mlanthology.org/neurips/2019/hsieh2019neurips-oneshot/}
}