A Unifying Bayesian Formulation of Measures of Interpretability in Human-AI

04/21/2021
by   Sarath Sreedharan, et al.
0

Existing approaches for generating human-aware agent behaviors have considered different measures of interpretability in isolation. Further, these measures have been studied under differing assumptions, thus precluding the possibility of designing a single framework that captures these measures under the same assumptions. In this paper, we present a unifying Bayesian framework that models a human observer's evolving beliefs about an agent and thereby define the problem of Generalized Human-Aware Planning. We will show that the definitions of interpretability measures like explicability, legibility and predictability from the prior literature fall out as special cases of our general framework. Through this framework, we also bring a previously ignored fact to light that the human-robot interactions are in effect open-world problems, particularly as a result of modeling the human's beliefs over the agent. Since the human may not only hold beliefs unknown to the agent but may also form new hypotheses about the agent when presented with novel or unexpected behaviors.

READ FULL TEXT
research
11/22/2020

A Bayesian Account of Measures of Interpretability in Human-AI Interaction

Existing approaches for the design of interpretable agent behavior consi...
research
07/06/2020

Dynamic Awareness

We investigate how to model the beliefs of an agent who becomes more awa...
research
09/11/2020

Mechanisms for a No-Regret Agent: Beyond the Common Prior

A rich class of mechanism design problems can be understood as incomplet...
research
12/27/2021

Interpreting Dynamical Systems as Bayesian Reasoners

A central concept in active inference is that the internal states of a p...
research
09/23/2022

Comparison of Lexical Alignment with a Teachable Robot in Human-Robot and Human-Human-Robot Interactions

Speakers build rapport in the process of aligning conversational behavio...
research
09/09/2019

Non-Bayesian Social Learning with Uncertain Models

Non-Bayesian social learning theory provides a framework that models dis...
research
08/12/2023

Latent Emission-Augmented Perspective-Taking (LEAPT) for Human-Robot Interaction

Perspective-taking is the ability to perceive or understand a situation ...

Please sign up or login with your details

Forgot password? Click here to reset