A Modern Self-Referential Weight Matrix That Learns to Modify Itself

Abstract

The weight matrix (WM) of a neural network (NN) is its program. The programs of many traditional NNs are learned through gradient descent in some error function, then remain fixed. The WM or program of a self-referential NN, however, can keep rapidly modifying all of itself during runtime. In principle, such NNs can meta-learn to learn, and meta-meta-learn to meta-learn to learn, and so on, in the sense of recursive self-improvement. Here we revisit such NNs, building upon recent successes of fast weight programmers (FWPs) and closely related linear Transformers. We propose a scalable self-referential WM (SRWM) that uses self-generated training patterns, outer products and the delta update rule to modify itself. We evaluate our SRWM in a multi-task reinforcement learning setting with procedurally generated ProcGen game environments. Our experiments demonstrate both practical applicability and competitive performance of the SRWM. Our code is public.

Cite

Text

Irie et al. "A Modern Self-Referential Weight Matrix That Learns to Modify Itself." NeurIPS 2021 Workshops: DeepRL, 2021.

Markdown

[Irie et al. "A Modern Self-Referential Weight Matrix That Learns to Modify Itself." NeurIPS 2021 Workshops: DeepRL, 2021.](https://mlanthology.org/neuripsw/2021/irie2021neuripsw-modern/)

BibTeX

@inproceedings{irie2021neuripsw-modern,
  title     = {{A Modern Self-Referential Weight Matrix That Learns to Modify Itself}},
  author    = {Irie, Kazuki and Schlag, Imanol and Csordás, Róbert and Schmidhuber, Jürgen},
  booktitle = {NeurIPS 2021 Workshops: DeepRL},
  year      = {2021},
  url       = {https://mlanthology.org/neuripsw/2021/irie2021neuripsw-modern/}
}