Contextual Priming and Feedback for Faster R-CNN

Abstract

The field of object detection has seen dramatic performance improvements in the last few years. Most of these gains are attributed to bottom-up, feedforward ConvNet frameworks. However, in case of humans, top-down information, context and feedback play an important role in doing object detection. This paper investigates how we can incorporate top-down information and feedback in the state-of-the-art Faster R-CNN framework. Specifically, we propose to: (a) augment Faster R-CNN with a semantic segmentation network; (b) use segmentation for top-down contextual priming; (c) use segmentation to provide top-down iterative feedback using two stage training. Our results indicate that all three contributions improve the performance on object detection, semantic segmentation and region proposal generation.

Cite

Text

Shrivastava and Gupta. "Contextual Priming and Feedback for Faster R-CNN." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46448-0_20

Markdown

[Shrivastava and Gupta. "Contextual Priming and Feedback for Faster R-CNN." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/shrivastava2016eccv-contextual/) doi:10.1007/978-3-319-46448-0_20

BibTeX

@inproceedings{shrivastava2016eccv-contextual,
  title     = {{Contextual Priming and Feedback for Faster R-CNN}},
  author    = {Shrivastava, Abhinav and Gupta, Abhinav},
  booktitle = {European Conference on Computer Vision},
  year      = {2016},
  pages     = {330-348},
  doi       = {10.1007/978-3-319-46448-0_20},
  url       = {https://mlanthology.org/eccv/2016/shrivastava2016eccv-contextual/}
}