Variable Computation in Recurrent Neural Networks

11/18/2016
by   Yacine Jernite, et al.
0

Recurrent neural networks (RNNs) have been used extensively and with increasing success to model various types of sequential data. Much of this progress has been achieved through devising recurrent units and architectures with the flexibility to capture complex statistics in the data, such as long range dependency or localized attention phenomena. However, while many sequential data (such as video, speech or language) can have highly variable information flow, most recurrent models still consume input features at a constant rate and perform a constant number of computations per time step, which can be detrimental to both speed and model capacity. In this paper, we explore a modification to existing recurrent units which allows them to learn to vary the amount of computation they perform at each step, without prior knowledge of the sequence's time structure. We show experimentally that not only do our models require fewer operations, they also lead to better performance overall on evaluation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/13/2022

EGRU: Event-based GRU for activity-sparse inference and learning

The scalability of recurrent neural networks (RNNs) is hindered by the s...
research
01/01/2015

Sequence Modeling using Gated Recurrent Neural Networks

In this paper, we have used Recurrent Neural Networks to capture and mod...
research
05/30/2019

A Lightweight Recurrent Network for Sequence Modeling

Recurrent networks have achieved great success on various sequential tas...
research
06/08/2015

Learning to Transduce with Unbounded Memory

Recently, strong results have been demonstrated by Deep Recurrent Neural...
research
08/15/2018

Using Regular Languages to Explore the Representational Capacity of Recurrent Neural Architectures

The presence of Long Distance Dependencies (LDDs) in sequential data pos...
research
09/21/2023

Parallelizing non-linear sequential models over the sequence length

Sequential models, such as Recurrent Neural Networks and Neural Ordinary...
research
03/29/2016

Adaptive Computation Time for Recurrent Neural Networks

This paper introduces Adaptive Computation Time (ACT), an algorithm that...

Please sign up or login with your details

Forgot password? Click here to reset