On the Power of Sigmoid Neural Networks

Abstract

We investigate the power of recurrent neural networks that apply the standard sigmoid activation function: a(z) = [2/(1 + e-”)] -1. We show that in the noiseless model, there exists a universal architecture that can be used to compute any recursive function. As a result, basic convergence questions concerning these architectures are shown to be undecidi~ble even for fixed-size networks. This is the first result of its kind for the standard sigmoid activation function; previous techniques only applied to linearized and truncated versions of this function. The significance of our result, besides the proving technique itself, lies in the popularity of the sigmoidal function both in applications of artificial neural networks and in models of biological neural networks. Our techniques can be applied to a much more general class of “sigmoid-like” activation functions, suggesting that Turing universality is a relatively common property of recurrent neural network models.

Cite

Text

Kilian and Siegelmann. "On the Power of Sigmoid Neural Networks." Annual Conference on Computational Learning Theory, 1993. doi:10.1145/168304.168321

Markdown

[Kilian and Siegelmann. "On the Power of Sigmoid Neural Networks." Annual Conference on Computational Learning Theory, 1993.](https://mlanthology.org/colt/1993/kilian1993colt-power/) doi:10.1145/168304.168321

BibTeX

@inproceedings{kilian1993colt-power,
  title     = {{On the Power of Sigmoid Neural Networks}},
  author    = {Kilian, Joe and Siegelmann, Hava T.},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {1993},
  pages     = {137-143},
  doi       = {10.1145/168304.168321},
  url       = {https://mlanthology.org/colt/1993/kilian1993colt-power/}
}