A Contraction Approach to Model-based Reinforcement Learning

09/18/2020
by   Ting-Han Fan, et al.
0

Model-based Reinforcement Learning has shown considerable experimental success. However, a theoretical understanding of it is still lacking. To this end, we analyze the error in cumulative reward for both stochastic and deterministic transitions using a contraction approach. We show that this approach doesn't require strong assumptions and can recover the typical quadratic error to the horizon. We prove that branched rollouts can reduce this error and are essential for deterministic transitions to have a Bellman contraction. Our results also apply to Imitation Learning, where we prove that GAN-type learning is better than Behavioral Cloning in continuous state and action spaces.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/26/2020

Episodic Self-Imitation Learning with Hindsight

Episodic self-imitation learning, a novel self-imitation algorithm with ...
research
02/19/2016

Policy Error Bounds for Model-Based Reinforcement Learning with Factored Linear Models

In this paper we study a model-based approach to calculating approximate...
research
11/16/2019

On Value Discrepancy of Imitation Learning

Imitation learning trains a policy from expert demonstrations. Imitation...
research
05/30/2019

Combating the Compounding-Error Problem with a Multi-step Model

Model-based reinforcement learning is an appealing framework for creatin...
research
04/20/2021

MBRL-Lib: A Modular Library for Model-based Reinforcement Learning

Model-based reinforcement learning is a compelling framework for data-ef...
research
09/25/2019

Model Imitation for Model-Based Reinforcement Learning

Model-based reinforcement learning (MBRL) aims to learn a dynamic model ...

Please sign up or login with your details

Forgot password? Click here to reset