STNet: Selective Tuning of Convolutional Networks for Object Localization

Abstract

Visual attention modeling has recently gained momentum in developing visual hierarchies provided by Convolutional Neural Networks. Despite recent successes of feed-forward processing on the Abstraction of concepts form raw images, the inherent nature of feedback processing has remained computationally controversial. Inspired by the computational models of covert visual attention, we propose the Selective Tuning of Convolutional Networks (STNet). It is composed of both streams of Bottom-Up and Top-Down information processing to selectively tune the visual representation of convolutional networks. We experimentally evaluate the performance of STNet for the weakly-supervised localization task on the ImageNet benchmark dataset. We demonstrate that STNet not only successfully surpasses the state-of-the-art results but also generates attention-driven class hypothesis maps.

Cite

Text

Biparva and Tsotsos. "STNet: Selective Tuning of Convolutional Networks for Object Localization." IEEE/CVF International Conference on Computer Vision Workshops, 2017. doi:10.1109/ICCVW.2017.319

Markdown

[Biparva and Tsotsos. "STNet: Selective Tuning of Convolutional Networks for Object Localization." IEEE/CVF International Conference on Computer Vision Workshops, 2017.](https://mlanthology.org/iccvw/2017/biparva2017iccvw-stnet/) doi:10.1109/ICCVW.2017.319

BibTeX

@inproceedings{biparva2017iccvw-stnet,
  title     = {{STNet: Selective Tuning of Convolutional Networks for Object Localization}},
  author    = {Biparva, Mahdi and Tsotsos, John K.},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2017},
  pages     = {2715-2723},
  doi       = {10.1109/ICCVW.2017.319},
  url       = {https://mlanthology.org/iccvw/2017/biparva2017iccvw-stnet/}
}