RRA: Recurrent Residual Attention for Sequence Learning

09/12/2017
by   Cheng Wang, et al.
0

In this paper, we propose a recurrent neural network (RNN) with residual attention (RRA) to learn long-range dependencies from sequential data. We propose to add residual connections across timesteps to RNN, which explicitly enhances the interaction between current state and hidden states that are several timesteps apart. This also allows training errors to be directly back-propagated through residual connections and effectively alleviates gradient vanishing problem. We further reformulate an attention mechanism over residual connections. An attention gate is defined to summarize the individual contribution from multiple previous hidden states in computing the current state. We evaluate RRA on three tasks: the adding problem, pixel-by-pixel MNIST classification and sentiment analysis on the IMDB dataset. Our experiments demonstrate that RRA yields better performance, faster convergence and more stable training compared to a standard LSTM network. Furthermore, RRA shows highly competitive performance to the state-of-the-art methods.

READ FULL TEXT
research
11/22/2016

Scene Labeling using Gated Recurrent Units with Explicit Long Range Conditioning

Recurrent neural network (RNN), as a powerful contextual dependency mode...
research
11/11/2019

Constructing Gradient Controllable Recurrent Neural Networks Using Hamiltonian Dynamics

Recurrent neural networks (RNNs) have gained a great deal of attention i...
research
04/07/2017

A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction

The Nonlinear autoregressive exogenous (NARX) model, which predicts the ...
research
09/27/2019

A Radio Signal Modulation Recognition Algorithm Based on Residual Networks and Attention Mechanisms

To solve the problem of inaccurate recognition of types of communication...
research
01/18/2018

Overcoming the vanishing gradient problem in plain recurrent networks

Plain recurrent networks greatly suffer from the vanishing gradient prob...
research
11/28/2018

Robust neural circuit reconstruction from serial electron microscopy with convolutional recurrent networks

Recent successes in deep learning have started to impact neuroscience. O...

Please sign up or login with your details

Forgot password? Click here to reset