The Uncanny Similarity of Recurrence and Depth
Abstract
It is widely believed that deep neural networks contain layer specialization, wherein networks extract hierarchical features representing edges and patterns in shallow layers and complete objects in deeper layers. Unlike common feed-forward models that have distinct filters at each layer, recurrent networks reuse the same parameters at various depths. In this work, we observe that recurrent models exhibit the same hierarchical behaviors and the same performance benefits as depth despite reusing the same filters at every recurrence. By training models of various feed-forward and recurrent architectures on several datasets for image classification as well as maze solving, we show that recurrent networks have the ability to closely emulate the behavior of non-recurrent deep models, often doing so with far fewer parameters.
Cite
Text
Schwarzschild et al. "The Uncanny Similarity of Recurrence and Depth." International Conference on Learning Representations, 2022.Markdown
[Schwarzschild et al. "The Uncanny Similarity of Recurrence and Depth." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/schwarzschild2022iclr-uncanny/)BibTeX
@inproceedings{schwarzschild2022iclr-uncanny,
title = {{The Uncanny Similarity of Recurrence and Depth}},
author = {Schwarzschild, Avi and Gupta, Arjun and Ghiasi, Amin and Goldblum, Micah and Goldstein, Tom},
booktitle = {International Conference on Learning Representations},
year = {2022},
url = {https://mlanthology.org/iclr/2022/schwarzschild2022iclr-uncanny/}
}