The Statistical Recurrent Unit

03/01/2017
by   Junier B. Oliva, et al.
0

Sophisticated gated recurrent neural network architectures like LSTMs and GRUs have been shown to be highly effective in a myriad of applications. We develop an un-gated unit, the statistical recurrent unit (SRU), that is able to learn long term dependencies in data by only keeping moving averages of statistics. The SRU's architecture is simple, un-gated, and contains a comparable number of parameters to LSTMs; yet, SRUs perform favorably to more sophisticated LSTM and GRU alternatives, often outperforming one or both in various tasks. We show the efficacy of SRUs as compared to LSTMs and GRUs in an unbiased manner by optimizing respective architectures' hyperparameters in a Bayesian optimization scheme for both synthetic and real-world tasks.

READ FULL TEXT
research
12/19/2016

A recurrent neural network without chaos

We introduce an exceptionally simple gated recurrent neural network (RNN...
research
03/23/2018

Can recurrent neural networks warp time?

Successful recurrent models such as long short-term memories (LSTMs) and...
research
03/05/2019

Gated Graph Convolutional Recurrent Neural Networks

Graph processes model a number of important problems such as identifying...
research
11/07/2017

Cortical microcircuits as gated-recurrent neural networks

Cortical circuits exhibit intricate recurrent architectures that are rem...
research
01/18/2018

Overcoming the vanishing gradient problem in plain recurrent networks

Plain recurrent networks greatly suffer from the vanishing gradient prob...
research
02/07/2017

Toward Abstraction from Multi-modal Data: Empirical Studies on Multiple Time-scale Recurrent Models

The abstraction tasks are challenging for multi- modal sequences as they...
research
09/03/2023

Traveling Waves Encode the Recent Past and Enhance Sequence Learning

Traveling waves of neural activity have been observed throughout the bra...

Please sign up or login with your details

Forgot password? Click here to reset