How Much Position Information Do Convolutional Neural Networks Encode?
Abstract
In contrast to fully connected networks, Convolutional Neural Networks (CNNs) achieve efficiency by learning weights associated with local filters with a finite spatial extent. An implication of this is that a filter may know what it is looking at, but not where it is positioned in the image. Information concerning absolute position is inherently useful, and it is reasonable to assume that deep CNNs may implicitly learn to encode this information if there is a means to do so. In this paper, we test this hypothesis revealing the surprising degree of absolute position information that is encoded in commonly used neural networks. A comprehensive set of experiments show the validity of this hypothesis and shed light on how and where this information is represented while offering clues to where positional information is derived from in deep CNNs.
Cite
Text
Islam et al. "How Much Position Information Do Convolutional Neural Networks Encode?." International Conference on Learning Representations, 2020.Markdown
[Islam et al. "How Much Position Information Do Convolutional Neural Networks Encode?." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/islam2020iclr-much/)BibTeX
@inproceedings{islam2020iclr-much,
title = {{How Much Position Information Do Convolutional Neural Networks Encode?}},
author = {Islam, Md Amirul and Jia, Sen and Bruce, Neil D. B.},
booktitle = {International Conference on Learning Representations},
year = {2020},
url = {https://mlanthology.org/iclr/2020/islam2020iclr-much/}
}