Multi-Scale Pyramid Pooling for Deep Convolutional Representation
Abstract
Compared to image representation based on low-level local descriptors, deep neural activations of Convolutional Neural Networks (CNNs) are richer in mid-level representation, but poorer in geometric invariance properties. In this paper, we present a straightforward framework for better image representation by combining the two approaches. To take advantages of both representations, we extract a fair amount of multi-scale dense local activations from a pre-trained CNN. We then aggregate the activations by Fisher kernel framework, which has been modified with a simple scale-wise normalization essential to make it suitable for CNN activations. Our representation demonstrates new state-of-the-art performances on three public datasets: 80.78% (Acc.) on MIT Indoor 67, 83.20% (mAP) on PASCAL VOC 2007 and 91.28% (Acc.) on Oxford 102 Flowers. The results suggest that our proposal can be used as a primary image representation for better performances in wide visual recognition tasks.
Cite
Text
Yoo et al. "Multi-Scale Pyramid Pooling for Deep Convolutional Representation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2015. doi:10.1109/CVPRW.2015.7301274Markdown
[Yoo et al. "Multi-Scale Pyramid Pooling for Deep Convolutional Representation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2015.](https://mlanthology.org/cvprw/2015/yoo2015cvprw-multiscale/) doi:10.1109/CVPRW.2015.7301274BibTeX
@inproceedings{yoo2015cvprw-multiscale,
title = {{Multi-Scale Pyramid Pooling for Deep Convolutional Representation}},
author = {Yoo, Donggeun and Park, Sunggyun and Lee, Joon-Young and Kweon, In So},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2015},
pages = {71-80},
doi = {10.1109/CVPRW.2015.7301274},
url = {https://mlanthology.org/cvprw/2015/yoo2015cvprw-multiscale/}
}