Gated recurrent units viewed through the lens of continuous time dynamical systems

06/03/2019
by   Ian D. Jordan, et al.
1

Gated recurrent units (GRUs) are specialized memory elements for building recurrent neural networks. Despite their incredible success in natural language, speech, and video processing, little is understood about the specific dynamics representable in a GRU network, along with the constraints these dynamics impose when generalizing a specific task. As a result, it is difficult to know a priori how successful a GRU network will perform on a given task. Using a continuous time analysis, we gain intuition on the inner workings of GRU networks. We restrict our presentation to low dimensions to allow for a comprehensive visualization. We found a surprisingly rich repertoire of dynamical features that includes stable limit cycles (nonlinear oscillations), multi-stable dynamics with various topologies, and homoclinic orbits. We contextualize the usefulness of the different kinds of dynamics and experimentally test their existence.

READ FULL TEXT

page 5

page 6

page 7

page 15

page 16

page 17

research
12/17/2015

Synthesis of recurrent neural networks for dynamical system simulation

We review several of the most widely used techniques for training recurr...
research
11/01/2018

Liquid Time-constant Recurrent Neural Networks as Universal Approximators

In this paper, we introduce the notion of liquid time-constant (LTC) rec...
research
08/10/2021

Recurrent neural network-based Internal Model Control of unknown nonlinear stable systems

Owing to their superior modeling capabilities, gated Recurrent Neural Ne...
research
06/23/2020

Lipschitz Recurrent Neural Networks

Differential equations are a natural choice for modeling recurrent neura...
research
05/08/2021

Pouring Dynamics Estimation Using Gated Recurrent Units

One of the most commonly performed manipulation in a human's daily life ...
research
10/11/2017

Discrete Event, Continuous Time RNNs

We investigate recurrent neural network architectures for event-sequence...
research
06/29/2006

May We Have Your Attention: Analysis of a Selective Attention Task

In this paper we present a deeper analysis than has previously been carr...

Please sign up or login with your details

Forgot password? Click here to reset