Compressed imitation learning

09/18/2020
by   Nathan Zhao, et al.
0

In analogy to compressed sensing, which allows sample-efficient signal reconstruction given prior knowledge of its sparsity in frequency domain, we propose to utilize policy simplicity (Occam's Razor) as a prior to enable sample-efficient imitation learning. We first demonstrated the feasibility of this scheme on linear case where state-value function can be sampled directly. We also extended the scheme to scenarios where only actions are visible and scenarios where the policy is obtained from nonlinear network. The method is benchmarked against behavior cloning and results in significantly higher scores with limited expert demonstrations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/16/2023

Sample-Efficient On-Policy Imitation Learning from Observations

Imitation learning from demonstrations (ILD) aims to alleviate numerous ...
research
05/23/2019

Action Assembly: Sparse Imitation Learning for Text Based Games with Combinatorial Action Spaces

We propose a computationally efficient algorithm that combines compresse...
research
06/20/2012

Imitation Learning with a Value-Based Prior

The goal of imitation learning is for an apprentice to learn how to beha...
research
06/26/2023

CEIL: Generalized Contextual Imitation Learning

In this paper, we present ContExtual Imitation Learning (CEIL), a genera...
research
01/21/2020

Loss-annealed GAIL for sample efficient and stable Imitation Learning

Imitation learning is the problem of learning a policy from an expert po...
research
12/23/2020

Augmenting Policy Learning with Routines Discovered from a Single Demonstration

Humans can abstract prior knowledge from very little data and use it to ...
research
12/02/2021

Quantile Filtered Imitation Learning

We introduce quantile filtered imitation learning (QFIL), a novel policy...

Please sign up or login with your details

Forgot password? Click here to reset