On Predictive Information Sub-optimality of RNNs

10/21/2019
by   Zhe Dong, et al.
0

Certain biological neurons demonstrate a remarkable capability to optimally compress the history of sensory inputs while being maximally informative about the future. In this work, we investigate if the same can be said of artificial neurons in recurrent neural networks (RNNs) trained with maximum likelihood. In experiments on two datasets, restorative Brownian motion and a hand-drawn sketch dataset, we find that RNNs are sub-optimal in the information plane. Instead of optimally compressing past information, they extract additional information that is not relevant for predicting the future. Overcoming this limitation may require alternative training procedures and architectures, or objectives beyond maximum likelihood estimation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2017

SEARNN: Training RNNs with Global-Local Losses

We propose SEARNN, a novel training algorithm for recurrent neural netwo...
research
06/09/2020

A bio-inspired bistable recurrent cell allows for long-lasting memory

Recurrent neural networks (RNNs) provide state-of-the-art performances i...
research
12/21/2019

Emergence of functional and structural properties of the head direction system by optimization of recurrent neural networks

Recent work suggests goal-driven training of neural networks can be used...
research
07/11/2021

Dual Training of Energy-Based Models with Overparametrized Shallow Neural Networks

Energy-based models (EBMs) are generative models that are usually traine...
research
05/23/2016

Path-Normalized Optimization of Recurrent Neural Networks with ReLU Activations

We investigate the parameter-space geometry of recurrent neural networks...
research
02/12/2019

Maximum Likelihood Estimation for Learning Populations of Parameters

Consider a setting with N independent individuals, each with an unknown ...
research
02/15/2017

Training Language Models Using Target-Propagation

While Truncated Back-Propagation through Time (BPTT) is the most popular...

Please sign up or login with your details

Forgot password? Click here to reset