Gated Linear Networks

Abstract

This paper presents a new family of backpropagation-free neural architectures, Gated Linear Networks (GLNs). What distinguishes GLNs from contemporary neural networks is the distributed and local nature of their credit assignment mechanism; each neuron directly predicts the target, forgoing the ability to learn feature representations in favor of rapid online learning. Individual neurons are able to model nonlinear functions via the use of data-dependent gating in conjunction with online convex optimization. We show that this architecture gives rise to universal learning capabilities in the limit, with effective model capacity increasing as a function of network size in a manner comparable with deep ReLU networks. Furthermore, we demonstrate that the GLN learning mechanism possesses extraordinary resilience to catastrophic forgetting, performing almost on par to an MLP with dropout and Elastic Weight Consolidation on standard benchmarks.

Cite

Text

Veness et al. "Gated Linear Networks." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I11.17202

Markdown

[Veness et al. "Gated Linear Networks." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/veness2021aaai-gated/) doi:10.1609/AAAI.V35I11.17202

BibTeX

@inproceedings{veness2021aaai-gated,
  title     = {{Gated Linear Networks}},
  author    = {Veness, Joel and Lattimore, Tor and Budden, David and Bhoopchand, Avishkar and Mattern, Christopher and Grabska-Barwinska, Agnieszka and Sezener, Eren and Wang, Jianan and Toth, Peter and Schmitt, Simon and Hutter, Marcus},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {10015-10023},
  doi       = {10.1609/AAAI.V35I11.17202},
  url       = {https://mlanthology.org/aaai/2021/veness2021aaai-gated/}
}