Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks

Abstract

In this paper we propose and investigate a novel nonlinear unit, called L p unit, for deep neural networks. The proposed L p unit receives signals from several projections of a subset of units in the layer below and computes a normalized L p norm. We notice two interesting interpretations of the L p unit. First, the proposed unit can be understood as a generalization of a number of conventional pooling operators such as average, root-mean-square and max pooling widely used in, for instance, convolutional neural networks (CNN), HMAX models and neocognitrons. Furthermore, the L p unit is, to a certain degree, similar to the recently proposed maxout unit [13] which achieved the state-of-the-art object recognition results on a number of benchmark datasets. Secondly, we provide a geometrical interpretation of the activation function based on which we argue that the L p unit is more efficient at representing complex, nonlinear separating boundaries. Each L p unit defines a superelliptic boundary, with its exact shape defined by the order p. We claim that this makes it possible to model arbitrarily shaped, curved boundaries more efficiently by combining a few L p units of different orders. This insight justifies the need for learning different orders for each unit in the model. We empirically evaluate the proposed L p units on a number of datasets and show that multilayer perceptrons (MLP) consisting of the L p units achieve the state-of-the-art results on a number of benchmark datasets. Furthermore, we evaluate the proposed L p unit on the recently proposed deep recurrent neural networks (RNN).

Cite

Text

Gülçehre et al. "Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2014. doi:10.1007/978-3-662-44848-9_34

Markdown

[Gülçehre et al. "Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2014.](https://mlanthology.org/ecmlpkdd/2014/gulcehre2014ecmlpkdd-learnednorm/) doi:10.1007/978-3-662-44848-9_34

BibTeX

@inproceedings{gulcehre2014ecmlpkdd-learnednorm,
  title     = {{Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks}},
  author    = {Gülçehre, Çaglar and Cho, KyungHyun and Pascanu, Razvan and Bengio, Yoshua},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2014},
  pages     = {530-546},
  doi       = {10.1007/978-3-662-44848-9_34},
  url       = {https://mlanthology.org/ecmlpkdd/2014/gulcehre2014ecmlpkdd-learnednorm/}
}