Dyna-AIL : Adversarial Imitation Learning by Planning

03/08/2019
by   Vaibhav Saxena, et al.
0

Adversarial methods for imitation learning have been shown to perform well on various control tasks. However, they require a large number of environment interactions for convergence. In this paper, we propose an end-to-end differentiable adversarial imitation learning algorithm in a Dyna-like framework for switching between model-based planning and model-free learning from expert data. Our results on both discrete and continuous environments show that our approach of using model-based planning along with model-free learning converges to an optimal policy with fewer number of environment interactions in comparison to the state-of-the-art learning methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/22/2020

Error Bounds of Imitating Policies and Environments

Imitation learning trains a policy by mimicking expert demonstrations. V...
research
04/03/2021

No Need for Interactions: Robust Model-Based Imitation Learning using Neural ODE

Interactions with either environments or expert policies during training...
research
09/06/2018

Sample-Efficient Imitation Learning via Generative Adversarial Nets

Recent work in imitation learning articulate their formulation around th...
research
09/29/2020

Learning Skills to Patch Plans Based on Inaccurate Models

Planners using accurate models can be effective for accomplishing manipu...
research
06/11/2017

Meta learning Framework for Automated Driving

The success of automated driving deployment is highly depending on the a...
research
05/27/2019

Provably Efficient Imitation Learning from Observation Alone

We study Imitation Learning (IL) from Observations alone (ILFO) in large...
research
04/02/2018

Universal Planning Networks

A key challenge in complex visuomotor control is learning abstract repre...

Please sign up or login with your details

Forgot password? Click here to reset