Applying Policy Iteration for Training Recurrent Neural Networks

10/02/2004
by   I. Szita, et al.
0

Recurrent neural networks are often used for learning time-series data. Based on a few assumptions we model this learning task as a minimization problem of a nonlinear least-squares cost function. The special structure of the cost function allows us to build a connection to reinforcement learning. We exploit this connection and derive a convergent, policy iteration-based algorithm. Furthermore, we argue that RNN training can be fit naturally into the reinforcement learning framework.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2018

Feature-Based Aggregation and Deep Reinforcement Learning: A Survey and Some New Implementations

In this paper we discuss policy iteration methods for approximate soluti...
research
08/01/2017

Tensorial Recurrent Neural Networks for Longitudinal Data Analysis

Traditional Recurrent Neural Networks assume vectorized data as inputs. ...
research
09/27/2016

Task Specific Adversarial Cost Function

The cost function used to train a generative model should fit the purpos...
research
12/17/2015

An Empirical Comparison of Neural Architectures for Reinforcement Learning in Partially Observable Environments

This paper explores the performance of fitted neural Q iteration for rei...
research
06/20/2019

The trade-off between long-term memory and smoothness for recurrent networks

Training recurrent neural networks (RNNs) that possess long-term memory ...
research
09/17/2018

Self Configuration in Machine Learning

In this paper we first present a class of algorithms for training multi-...

Please sign up or login with your details

Forgot password? Click here to reset