ContextLocNet: Context-Aware Deep Network Models for Weakly Supervised Localization
Abstract
We aim to localize objects in images using image-level supervision only. Previous approaches to this problem mainly focus on discriminative object regions and often fail to locate precise object boundaries. We address this problem by introducing two types of context-aware guidance models, additive and contrastive models, that leverage their surrounding context regions to improve localization. The additive model encourages the predicted object region to be supported by its surrounding context region. The contrastive model encourages the predicted object region to be outstanding from its surrounding context region. Our approach benefits from the recent success of convolutional neural networks for object recognition and extends Fast R-CNN to weakly supervised object localization. Extensive experimental evaluation on the PASCAL VOC 2007 and 2012 benchmarks shows that our context-aware approach significantly improves weakly supervised localization and detection.
Cite
Text
Kantorov et al. "ContextLocNet: Context-Aware Deep Network Models for Weakly Supervised Localization." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46454-1_22Markdown
[Kantorov et al. "ContextLocNet: Context-Aware Deep Network Models for Weakly Supervised Localization." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/kantorov2016eccv-contextlocnet/) doi:10.1007/978-3-319-46454-1_22BibTeX
@inproceedings{kantorov2016eccv-contextlocnet,
title = {{ContextLocNet: Context-Aware Deep Network Models for Weakly Supervised Localization}},
author = {Kantorov, Vadim and Oquab, Maxime and Cho, Minsu and Laptev, Ivan},
booktitle = {European Conference on Computer Vision},
year = {2016},
pages = {350-365},
doi = {10.1007/978-3-319-46454-1_22},
url = {https://mlanthology.org/eccv/2016/kantorov2016eccv-contextlocnet/}
}