Input-Output Equivalence of Unitary and Contractive RNNs

10/30/2019
by   M. Emami, et al.
0

Unitary recurrent neural networks (URNNs) have been proposed as a method to overcome the vanishing and exploding gradient problem in modeling data with long-term dependencies. A basic question is how restrictive is the unitary constraint on the possible input-output mappings of such a network? This work shows that for any contractive RNN with ReLU activations, there is a URNN with at most twice the number of hidden states and the identical input-output mapping. Hence, with ReLU activations, URNNs are as expressive as general RNNs. In contrast, for certain smooth activations, it is shown that the input-output mapping of an RNN cannot be matched with a URNN, even with an arbitrary number of states. The theoretical results are supported by experiments on modeling of slowly-varying dynamical systems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/06/2021

Metric Entropy Limits on Recurrent Neural Network Learning of Linear Dynamical Systems

One of the most influential results in neural network theory is the univ...
research
06/21/2019

Universal Approximation of Input-Output Maps by Temporal Convolutional Nets

There has been a recent shift in sequence-to-sequence modeling from recu...
research
10/03/2022

Random orthogonal additive filters: a solution to the vanishing/exploding gradient of deep neural networks

Since the recognition in the early nineties of the vanishing/exploding (...
research
05/23/2016

Path-Normalized Optimization of Recurrent Neural Networks with ReLU Activations

We investigate the parameter-space geometry of recurrent neural networks...
research
05/21/2018

Learning Device Models with Recurrent Neural Networks

Recurrent neural networks (RNNs) are powerful constructs capable of mode...
research
02/13/2023

Fourier-RNNs for Modelling Noisy Physics Data

Classical sequential models employed in time-series prediction rely on l...
research
10/08/2019

Inferring Dynamical Systems with Long-Range Dependencies through Line Attractor Regularization

Vanilla RNN with ReLU activation have a simple structure that is amenabl...

Please sign up or login with your details

Forgot password? Click here to reset