Learning with Partial Forgetting in Modern Hopfield Networks

Abstract

It has been known by neuroscience studies that partial and transient forgetting of memory often plays an important role in the brain to improve performance for certain intellectual activities. In machine learning, associative memory models such as classical and modern Hopfield networks have been proposed to express memories as attractors in the feature space of a closed recurrent network. In this work, we propose learning with partial forgetting (LwPF), where a partial forgetting functionality is designed by element-wise non-bijective projections, for memory neurons in modern Hopfield networks to improve model performance. We incorporate LwPF into the attention mechanism also, whose process has been shown to be identical to the update rule of a certain modern Hopfield network, by modifying the corresponding Lagrangian. We evaluated the effectiveness of LwPF on three diverse tasks such as bit-pattern classification, immune repertoire classification for computational biology, and image classification for computer vision, and confirmed that LwPF consistently improves the performance of existing neural networks including DeepRC and vision transformers.

Cite

Text

Ota et al. "Learning with Partial Forgetting in Modern Hopfield Networks." Artificial Intelligence and Statistics, 2023.

Markdown

[Ota et al. "Learning with Partial Forgetting in Modern Hopfield Networks." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/ota2023aistats-learning/)

BibTeX

@inproceedings{ota2023aistats-learning,
  title     = {{Learning with Partial Forgetting in Modern Hopfield Networks}},
  author    = {Ota, Toshihiro and Sato, Ikuro and Kawakami, Rei and Tanaka, Masayuki and Inoue, Nakamasa},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2023},
  pages     = {6661-6673},
  volume    = {206},
  url       = {https://mlanthology.org/aistats/2023/ota2023aistats-learning/}
}