| Article ID | Journal | Published Year | Pages | File Type | 
|---|---|---|---|---|
| 10146098 | Pattern Recognition Letters | 2018 | 11 Pages | 
Abstract
												We independently reproduce the QRNN experiments of Bradbury et al. [1] and compare our DReLU-based QRNNs with the original tanh-based QRNNs and Long Short-Term Memory networks (LSTMs) on sentiment classification and word-level language modeling. Additionally, we evaluate on character-level language modeling, showing that we are able to stack up to eight QRNN layers with DReLUs, thus making it possible to improve the current state-of-the-art in character-level language modeling over shallow architectures based on LSTMs.
											Related Topics
												
													Physical Sciences and Engineering
													Computer Science
													Computer Vision and Pattern Recognition
												
											Authors
												Fréderic Godin, Jonas Degrave, Joni Dambre, Wesley De Neve, 
											