Introducing CURRENNT: the Munich Open-Source CUDA RecurREnt Neural Network Toolkit
File(s)weninger15a.pdf (285.94 KB)
Published version
Author(s)
Weninger, F
Bergmann, J
Schuller, B
Type
Journal Article
Abstract
In this article, we introduce CURRENNT, an open-source parallel implementation of deep recurrent neural networks (RNNs) supporting graphics processing units (GPUs) through NVIDIA's Computed Unified Device Architecture (CUDA). CURRENNT supports uni- and bidirectional RNNs with Long Short-Term Memory (LSTM) memory cells which overcome the vanishing gradient problem. To our knowledge, CURRENNT is the first publicly available parallel implementation of deep LSTM-RNNs. Benchmarks are given on a noisy speech recognition task from the 2013 2nd CHiME Speech Separation and Recognition Challenge, where LSTM-RNNs have been shown to deliver best performance. In the result, double digit speedups in bidirectional LSTM training are achieved with respect to a reference single-threaded CPU implementation. CURRENNT is available under the GNU General Public License from http://sourceforge.net/p/currennt.
Date Issued
2015-03-31
Date Acceptance
2013-07-01
Citation
Journal of Machine Learning Research, 2015, 16, pp.547-551
ISSN
1532-4435
Publisher
Journal of Machine Learning Research
Start Page
547
End Page
551
Journal / Book Title
Journal of Machine Learning Research
Volume
16
Copyright Statement
© The Author(s) 2015.
Subjects
Science & Technology
Technology
Automation & Control Systems
Computer Science, Artificial Intelligence
Computer Science
parallel computing
deep neural networks
recurrent neural networks
Long Short-Term Memory
Artificial Intelligence & Image Processing
08 Information And Computing Sciences
17 Psychology And Cognitive Sciences
Notes
5 pages, (acceptance rate: 18\,%, IF: 2.853, 5-year IF: 4.649 (2013))
Publication Status
Published