Robust Inverse Reinforcement Learning under Transition Dynamics Mismatch

07/02/2020
by   Luca Viano, et al.
0

We study the inverse reinforcement learning (IRL) problem under the transition dynamics mismatch between the expert and the learner. In particular, we consider the Maximum Causal Entropy (MCE) IRL learner model and provide an upper bound on the learner's performance degradation based on the ℓ_1-distance between the two transition dynamics of the expert and the learner. Then, by leveraging insights from the Robust RL literature, we propose a robust MCE IRL algorithm, which is a principled approach to help with this mismatch issue. Finally, we empirically demonstrate the stable performance of our algorithm compared to the standard MCE IRL algorithm under transition mismatches in finite MDP problems.

READ FULL TEXT

page 8

page 31

page 33

page 34

page 35

research
04/25/2022

Imitation Learning from Observations under Transition Model Disparity

Learning to perform tasks by leveraging a dataset of expert observations...
research
10/07/2020

Regularized Inverse Reinforcement Learning

Inverse Reinforcement Learning (IRL) aims to facilitate a learner's abil...
research
10/27/2017

Inverse Reinforcement Learning Under Noisy Observations

We consider the problem of performing inverse reinforcement learning whe...
research
01/05/2023

Data-Driven Inverse Reinforcement Learning for Expert-Learner Zero-Sum Games

In this paper, we formulate inverse reinforcement learning (IRL) as an e...
research
10/18/2021

Provable Hierarchy-Based Meta-Reinforcement Learning

Hierarchical reinforcement learning (HRL) has seen widespread interest a...
research
06/02/2022

Uniqueness and Complexity of Inverse MDP Models

What is the action sequence aa'a" that was likely responsible for reachi...
research
02/11/2020

Objective Mismatch in Model-based Reinforcement Learning

Model-based reinforcement learning (MBRL) has been shown to be a powerfu...

Please sign up or login with your details

Forgot password? Click here to reset