Attention and Augmented Recurrent Neural Networks
Abstract
Distill articles are interactive publications and do not include traditional abstracts. This summary was written for the ML Anthology. Explores four extensions to recurrent neural networks—neural Turing machines, attentional interfaces, adaptive computation time, and neural programmers—unified by attention as a mechanism for focusing computation on relevant information.
Cite
Text
Olah and Carter. "Attention and Augmented Recurrent Neural Networks." Distill, 2016. doi:10.23915/distill.00001Markdown
[Olah and Carter. "Attention and Augmented Recurrent Neural Networks." Distill, 2016.](https://mlanthology.org/distill/2016/olah2016distill-attention/) doi:10.23915/distill.00001BibTeX
@article{olah2016distill-attention,
title = {{Attention and Augmented Recurrent Neural Networks}},
author = {Olah, Chris and Carter, Shan},
journal = {Distill},
year = {2016},
doi = {10.23915/distill.00001},
url = {https://mlanthology.org/distill/2016/olah2016distill-attention/}
}