Distal Explanations for Explainable Reinforcement Learning Agents

01/28/2020
by   Prashan Madumal, et al.
14

Causal explanations present an intuitive way to understand the course of events through causal chains, and are widely accepted in cognitive science as the prominent model humans use for explanation. Importantly, causal models can generate opportunity chains, which take the form of `A enables B and B causes C'. We ground the notion of opportunity chains in human-agent experimental data, where we present participants with explanations from different models and ask them to provide their own explanations for agent behaviour. Results indicate that humans do in-fact use the concept of opportunity chains frequently for describing artificial agent behaviour. Recently, action influence models have been proposed to provide causal explanations for model-free reinforcement learning (RL). While these models can generate counterfactuals—things that did not happen but could have under different conditions—they lack the ability to generate explanations of opportunity chains. We introduce a distal explanation model that can analyse counterfactuals and opportunity chains using decision trees and causal models. We employ a recurrent neural network to learn opportunity chains and make use of decision trees to improve the accuracy of task prediction and the generated counterfactuals. We computationally evaluate the model in 6 RL benchmarks using different RL algorithms, and show that our model performs better in task prediction. We report on a study with 90 participants who receive explanations of RL agents behaviour in solving three scenarios: 1) Adversarial; 2) Search and rescue; and 3) Human-Agent collaborative scenarios. We investigate the participants' understanding of the agent through task prediction and their subjective satisfaction of the explanations and show that our distal explanation model results in improved outcomes over the three scenarios compared with two baseline explanation models.

READ FULL TEXT
research
05/27/2019

Explainable Reinforcement Learning Through a Causal Lens

Prevalent theories in cognitive science propose that humans understand a...
research
03/09/2023

Explainable Goal Recognition: A Framework Based on Weight of Evidence

We introduce and evaluate an eXplainable Goal Recognition (XGR) model th...
research
01/29/2021

Counterfactual State Explanations for Reinforcement Learning Agents via Generative Deep Learning

Counterfactual explanations, which deal with "why not?" scenarios, can p...
research
03/05/2021

Causal Analysis of Agent Behavior for AI Safety

As machine learning systems become more powerful they also become increa...
research
04/01/2023

Conveying Autonomous Robot Capabilities through Contrasting Behaviour Summaries

As advances in artificial intelligence enable increasingly capable learn...
research
11/10/2020

What Did You Think Would Happen? Explaining Agent Behaviour Through Intended Outcomes

We present a novel form of explanation for Reinforcement Learning, based...
research
12/01/2022

Decisions that Explain Themselves: A User-Centric Deep Reinforcement Learning Explanation System

With deep reinforcement learning (RL) systems like autonomous driving be...

Please sign up or login with your details

Forgot password? Click here to reset