On the Memory Mechanism of Tensor-Power Recurrent Models

03/02/2021
by   Hejia Qiu, et al.
0

Tensor-power (TP) recurrent model is a family of non-linear dynamical systems, of which the recurrence relation consists of a p-fold (a.k.a., degree-p) tensor product. Despite such the model frequently appears in the advanced recurrent neural networks (RNNs), to this date there is limited study on its memory property, a critical characteristic in sequence tasks. In this work, we conduct a thorough investigation of the memory mechanism of TP recurrent models. Theoretically, we prove that a large degree p is an essential condition to achieve the long memory effect, yet it would lead to unstable dynamical behaviors. Empirically, we tackle this issue by extending the degree p from discrete to a differentiable domain, such that it is efficiently learnable from a variety of datasets. Taken together, the new model is expected to benefit from the long memory effect in a stable manner. We experimentally show that the proposed model achieves competitive performance compared to various advanced RNNs in both the single-cell and seq2seq architectures.

READ FULL TEXT
research
09/16/2020

On the Curse of Memory in Recurrent Neural Networks: Approximation and Optimization Analysis

We study the approximation properties and optimization dynamics of recur...
research
04/11/2020

Convex Sets of Robust Recurrent Neural Networks

Recurrent neural networks (RNNs) are a class of nonlinear dynamical syst...
research
05/25/2017

Predictive State Recurrent Neural Networks

We present a new model, Predictive State Recurrent Neural Networks (PSRN...
research
10/28/2020

The geometry of integration in text classification RNNs

Despite the widespread application of recurrent neural networks (RNNs) a...
research
06/14/2018

Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks

Recurrent neural networks have gained widespread use in modeling sequenc...
research
10/08/2020

A Fully Tensorized Recurrent Neural Network

Recurrent neural networks (RNNs) are powerful tools for sequential model...
research
01/08/2021

Slow manifolds in recurrent networks encode working memory efficiently and robustly

Working memory is a cognitive function involving the storage and manipul...

Please sign up or login with your details

Forgot password? Click here to reset