General Value Function Networks

07/18/2018
by   Matthew Schlegel, et al.
0

In this paper we show that restricting the representation-layer of a Recurrent Neural Network (RNN) improves accuracy and reduces the depth of recursive training procedures in partially observable domains. Artificial Neural Networks have been shown to learn useful state representations for high-dimensional visual and continuous control domains. If the the tasks at hand exhibits long depends back in time, these instantaneous feed-forward approaches are augmented with recurrent connections and trained with Back-prop Through Time (BPTT). This unrolled training can become computationally prohibitive if the dependency structure is long, and while recent work on LSTMs and GRUs has improved upon naive training strategies, there is still room for improvements in computational efficiency and parameter sensitivity. In this paper we explore a simple modification to the classic RNN structure: restricting the state to be comprised of multi-step General Value Function predictions. We formulate an architecture called General Value Function Networks (GVFNs), and corresponding objective that generalizes beyond previous approaches. We show that our GVFNs are significantly more robust to train, and facilitate accurate prediction with no gradients needed back-in-time in domains with substantial long-term dependences.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/06/2018

Residual Memory Networks: Feed-forward approach to learn long temporal dependencies

Training deep recurrent neural network (RNN) architectures is complicate...
research
05/23/2019

Recurrent Value Functions

Despite recent successes in Reinforcement Learning, value-based methods ...
research
01/30/2017

Memory Augmented Neural Networks with Wormhole Connections

Recent empirical results on long-term dependency tasks have shown that n...
research
07/28/2023

Dynamic Analysis and an Eigen Initializer for Recurrent Neural Networks

In recurrent neural networks, learning long-term dependency is the main ...
research
04/18/2019

Interplanetary Transfers via Deep Representations of the Optimal Policy and/or of the Value Function

A number of applications to interplanetary trajectories have been recent...
research
04/14/2016

Learning Visual Storylines with Skipping Recurrent Neural Networks

What does a typical visit to Paris look like? Do people first take photo...
research
05/04/2020

Neural Networks and Value at Risk

Utilizing a generative regime switching framework, we perform Monte-Carl...

Please sign up or login with your details

Forgot password? Click here to reset