Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
10146098 | Pattern Recognition Letters | 2018 | 11 Pages |
Abstract
We independently reproduce the QRNN experiments of Bradbury et al. [1] and compare our DReLU-based QRNNs with the original tanh-based QRNNs and Long Short-Term Memory networks (LSTMs) on sentiment classification and word-level language modeling. Additionally, we evaluate on character-level language modeling, showing that we are able to stack up to eight QRNN layers with DReLUs, thus making it possible to improve the current state-of-the-art in character-level language modeling over shallow architectures based on LSTMs.
Related Topics
Physical Sciences and Engineering
Computer Science
Computer Vision and Pattern Recognition
Authors
Fréderic Godin, Jonas Degrave, Joni Dambre, Wesley De Neve,