Quasi-Recurrent Neural Networks

11/05/2016
by   James Bradbury, et al.
0

Recurrent neural networks are a powerful tool for modeling sequential data, but the dependence of each timestep's computation on the previous timestep's output limits parallelism and makes RNNs unwieldy for very long sequences. We introduce quasi-recurrent neural networks (QRNNs), an approach to neural sequence modeling that alternates convolutional layers, which apply in parallel across timesteps, and a minimalist recurrent pooling function that applies in parallel across channels. Despite lacking trainable recurrent layers, stacked QRNNs have better predictive accuracy than stacked LSTMs of the same hidden size. Due to their increased parallelism, they are up to 16 times faster at train and test time. Experiments on language modeling, sentiment classification, and character-level neural machine translation demonstrate these advantages and underline the viability of QRNNs as a basic building block for a variety of sequence tasks.

READ FULL TEXT
research
09/08/2014

Recurrent Neural Network Regularization

We present a simple regularization technique for Recurrent Neural Networ...
research
07/25/2017

Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation Functions in Quasi-Recurrent Neural Networks

In this paper, we introduce a novel type of Rectified Linear Unit (ReLU)...
research
07/19/2017

Improving Language Modeling using Densely Connected Recurrent Neural Networks

In this paper, we introduce the novel concept of densely connected layer...
research
05/17/2016

Deep Action Sequence Learning for Causal Shape Transformation

Deep learning became the method of choice in recent year for solving a w...
research
11/19/2019

Thick-Net: Parallel Network Structure for Sequential Modeling

Recurrent neural networks have been widely used in sequence learning tas...
research
09/01/2018

Finding the Answers with Definition Models

Inspired by a previous attempt to answer crossword questions using neura...
research
09/14/2023

Advancing Regular Language Reasoning in Linear Recurrent Neural Networks

In recent studies, linear recurrent neural networks (LRNNs) have achieve...

Please sign up or login with your details

Forgot password? Click here to reset