Neural Speed Reading via Skim-RNN
Abstract
Inspired by the principles of speed reading, we introduce Skim-RNN, a recurrent neural network (RNN) that dynamically decides to update only a small fraction of the hidden state for relatively unimportant input tokens. Skim-RNN gives a significant computational advantage over an RNN that always updates the entire hidden state. Skim-RNN uses the same input and output interfaces as a standard RNN and can be easily used instead of RNNs in existing models. In our experiments, we show that Skim-RNN can achieve significantly reduced computational cost without losing accuracy compared to standard RNNs across five different natural language tasks. In addition, we demonstrate that the trade-off between accuracy and speed of Skim-RNN can be dynamically controlled during inference time in a stable manner. Our analysis also shows that Skim-RNN running on a single CPU offers lower latency compared to standard RNNs on GPUs.
Cite
Text
Seo et al. "Neural Speed Reading via Skim-RNN." International Conference on Learning Representations, 2018.Markdown
[Seo et al. "Neural Speed Reading via Skim-RNN." International Conference on Learning Representations, 2018.](https://mlanthology.org/iclr/2018/seo2018iclr-neural/)BibTeX
@inproceedings{seo2018iclr-neural,
title = {{Neural Speed Reading via Skim-RNN}},
author = {Seo, Minjoon and Min, Sewon and Farhadi, Ali and Hajishirzi, Hannaneh},
booktitle = {International Conference on Learning Representations},
year = {2018},
url = {https://mlanthology.org/iclr/2018/seo2018iclr-neural/}
}