Adaptive Bases for Reinforcement Learning

05/02/2010
by   Dotan Di Castro, et al.
0

We consider the problem of reinforcement learning using function approximation, where the approximating basis can change dynamically while interacting with the environment. A motivation for such an approach is maximizing the value function fitness to the problem faced. Three errors are considered: approximation square error, Bellman residual, and projected Bellman residual. Algorithms under the actor-critic framework are presented, and shown to converge. The advantage of such an adaptive basis is demonstrated in simulations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/06/2021

Error Controlled Actor-Critic

On error of value function inevitably causes an overestimation phenomeno...
research
08/03/2021

Variational Actor-Critic Algorithms

We introduce a class of variational actor-critic algorithms based on a v...
research
09/16/2009

A Convergent Online Single Time Scale Actor Critic Algorithm

Actor-Critic based approaches were among the first to address reinforcem...
research
08/22/2019

On Convergence Rate of Adaptive Multiscale Value Function Approximation For Reinforcement Learning

In this paper, we propose a generic framework for devising an adaptive a...
research
02/22/2021

Actor-Critic Method for High Dimensional Static Hamilton–Jacobi–Bellman Partial Differential Equations based on Neural Networks

We propose a novel numerical method for high dimensional Hamilton–Jacobi...
research
04/22/2022

Adaptive Online Value Function Approximation with Wavelets

Using function approximation to represent a value function is necessary ...
research
06/03/2011

Efficient Reinforcement Learning Using Recursive Least-Squares Methods

The recursive least-squares (RLS) algorithm is one of the most well-know...

Please sign up or login with your details

Forgot password? Click here to reset