"Guess what I'm doing": Extending legibility to sequential decision tasks

09/19/2022
by   Miguel Faria, et al.
2

In this paper we investigate the notion of legibility in sequential decision tasks under uncertainty. Previous works that extend legibility to scenarios beyond robot motion either focus on deterministic settings or are computationally too expensive. Our proposed approach, dubbed PoL-MDP, is able to handle uncertainty while remaining computationally tractable. We establish the advantages of our approach against state-of-the-art approaches in several simulated scenarios of different complexity. We also showcase the use of our legible policies as demonstrations for an inverse reinforcement learning agent, establishing their superiority against the commonly used demonstrations based on the optimal policy. Finally, we assess the legibility of our computed policies through a user study where people are asked to infer the goal of a mobile robot following a legible policy by observing its actions.

READ FULL TEXT
research
10/28/2021

Learning Feasibility to Imitate Demonstrators with Different Dynamics

The goal of learning from demonstrations is to learn a policy for an age...
research
02/15/2021

Learning from Demonstrations using Signal Temporal Logic

Learning-from-demonstrations is an emerging paradigm to obtain effective...
research
03/23/2023

Boosting Reinforcement Learning and Planning with Demonstrations: A Survey

Although reinforcement learning has seen tremendous success recently, th...
research
11/15/2022

PARTNR: Pick and place Ambiguity Resolving by Trustworthy iNteractive leaRning

Several recent works show impressive results in mapping language-based h...
research
05/20/2018

Machine Teaching for Inverse Reinforcement Learning: Algorithms and Applications

Inverse reinforcement learning (IRL) infers a reward function from demon...
research
07/18/2019

Composing Diverse Policies for Temporally Extended Tasks

Temporally extended and sequenced robot motion tasks are often character...
research
08/09/2023

Bayesian Inverse Transition Learning for Offline Settings

Offline Reinforcement learning is commonly used for sequential decision-...

Please sign up or login with your details

Forgot password? Click here to reset