Non-normal Recurrent Neural Network (nnRNN): learning long time dependencies while improving expressivity with transient dynamics

05/28/2019
by   Giancarlo Kerg, et al.
1

A recent strategy to circumvent the exploding and vanishing gradient problem in RNNs, and to allow the stable propagation of signals over long time scales, is to constrain recurrent connectivity matrices to be orthogonal or unitary. This ensures eigenvalues with unit norm and thus stable dynamics and training. However this comes at the cost of reduced expressivity due to the limited variety of orthogonal transformations. We propose a novel connectivity structure based on the Schur decomposition and a splitting of the Schur form into normal and non-normal parts. This allows to parametrize matrices with unit-norm eigenspectra without orthogonality constraints on eigenbases. The resulting architecture ensures access to a larger space of spectrally constrained matrices, of which orthogonal matrices are a subset. This crucial difference retains the stability advantages and training speed of orthogonal RNNs while enhancing expressivity, especially on tasks that require computations over ongoing input sequences.

READ FULL TEXT

page 4

page 5

page 7

page 9

page 10

page 12

page 13

page 14

research
05/31/2019

Improved memory in recurrent neural networks with sequential non-normal dynamics

Training recurrent neural networks (RNNs) is a hard problem due to degen...
research
08/12/2022

Orthogonal Gated Recurrent Unit with Neumann-Cayley Transformation

In recent years, using orthogonal matrices has been shown to be a promis...
research
11/18/2019

Eigenvalue Normalized Recurrent Neural Networks for Short Term Memory

Several variants of recurrent neural networks (RNNs) with orthogonal or ...
research
02/14/2023

Convolutional unitary or orthogonal recurrent neural networks

Recurrent neural networks are extremely powerful yet hard to train. One ...
research
04/24/2023

Adaptive-saturated RNN: Remember more with less instability

Orthogonal parameterization is a compelling solution to the vanishing gr...
research
09/07/2023

Brief technical note on linearizing recurrent neural networks (RNNs) before vs after the pointwise nonlinearity

Linearization of the dynamics of recurrent neural networks (RNNs) is oft...
research
01/24/2019

Cheap Orthogonal Constraints in Neural Networks: A Simple Parametrization of the Orthogonal and Unitary Group

We introduce a novel approach to perform first-order optimization with o...

Please sign up or login with your details

Forgot password? Click here to reset