Introducing CURRENNT: The Munich Open-Source CUDA RecurREnt Neural Network Toolkit
Abstract
In this article, we introduce CURRENNT, an open-source parallel implementation of deep recurrent neural networks (RNNs) supporting graphics processing units (GPUs) through NVIDIA's Computed Unified Device Architecture (CUDA). CURRENNT supports uni- and bidirectional RNNs with Long Short-Term Memory (LSTM) memory cells which overcome the vanishing gradient problem. To our knowledge, CURRENNT is the first publicly available parallel implementation of deep LSTM-RNNs. Benchmarks are given on a noisy speech recognition task from the 2013 2nd CHiME Speech Separation and Recognition Challenge, where LSTM-RNNs have been shown to deliver best performance. In the result, double digit speedups in bidirectional LSTM training are achieved with respect to a reference single-threaded CPU implementation. CURRENNT is available under the GNU General Public License from http://sourceforge.net/p/currennt.
Cite
Text
Weninger. "Introducing CURRENNT: The Munich Open-Source CUDA RecurREnt Neural Network Toolkit." Machine Learning Open Source Software, 2015.Markdown
[Weninger. "Introducing CURRENNT: The Munich Open-Source CUDA RecurREnt Neural Network Toolkit." Machine Learning Open Source Software, 2015.](https://mlanthology.org/mloss/2015/weninger2015jmlr-introducing/)BibTeX
@article{weninger2015jmlr-introducing,
title = {{Introducing CURRENNT: The Munich Open-Source CUDA RecurREnt Neural Network Toolkit}},
author = {Weninger, Felix},
journal = {Machine Learning Open Source Software},
year = {2015},
pages = {547-551},
volume = {16},
url = {https://mlanthology.org/mloss/2015/weninger2015jmlr-introducing/}
}