Attention and Augmented Recurrent Neural Networks

Abstract

Distill articles are interactive publications and do not include traditional abstracts. This summary was written for the ML Anthology. Explores four extensions to recurrent neural networks—neural Turing machines, attentional interfaces, adaptive computation time, and neural programmers—unified by attention as a mechanism for focusing computation on relevant information.

Cite

Text

Olah and Carter. "Attention and Augmented Recurrent Neural Networks." Distill, 2016. doi:10.23915/distill.00001

Markdown

[Olah and Carter. "Attention and Augmented Recurrent Neural Networks." Distill, 2016.](https://mlanthology.org/distill/2016/olah2016distill-attention/) doi:10.23915/distill.00001

BibTeX

@article{olah2016distill-attention,
  title     = {{Attention and Augmented Recurrent Neural Networks}},
  author    = {Olah, Chris and Carter, Shan},
  journal   = {Distill},
  year      = {2016},
  doi       = {10.23915/distill.00001},
  url       = {https://mlanthology.org/distill/2016/olah2016distill-attention/}
}