One-Shot Object Detection with Co-Attention and Co-Excitation
Abstract
This paper aims to tackle the challenging problem of one-shot object detection. Given a query image patch whose class label is not included in the training data, the goal of the task is to detect all instances of the same class in a target image. To this end, we develop a novel {\em co-attention and co-excitation} (CoAE) framework that makes contributions in three key technical aspects. First, we propose to use the non-local operation to explore the co-attention embodied in each query-target pair and yield region proposals accounting for the one-shot situation. Second, we formulate a squeeze-and-co-excitation scheme that can adaptively emphasize correlated feature channels to help uncover relevant proposals and eventually the target objects. Third, we design a margin-based ranking loss for implicitly learning a metric to predict the similarity of a region proposal to the underlying query, no matter its class label is seen or unseen in training. The resulting model is therefore a two-stage detector that yields a strong baseline on both VOC and MS-COCO under one-shot setting of detecting objects from both seen and never-seen classes.
Cite
Text
Hsieh et al. "One-Shot Object Detection with Co-Attention and Co-Excitation." Neural Information Processing Systems, 2019.Markdown
[Hsieh et al. "One-Shot Object Detection with Co-Attention and Co-Excitation." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/hsieh2019neurips-oneshot/)BibTeX
@inproceedings{hsieh2019neurips-oneshot,
title = {{One-Shot Object Detection with Co-Attention and Co-Excitation}},
author = {Hsieh, Ting-I and Lo, Yi-Chen and Chen, Hwann-Tzong and Liu, Tyng-Luh},
booktitle = {Neural Information Processing Systems},
year = {2019},
pages = {2725-2734},
url = {https://mlanthology.org/neurips/2019/hsieh2019neurips-oneshot/}
}