Local and Global Explanations of Agent Behavior: Integrating Strategy Summaries with Saliency Maps

05/18/2020
by   Tobias Huber, et al.
9

With advances in reinforcement learning (RL), agents are now being developed in high-stakes application domains such as healthcare and transportation. Explaining the behavior of these agents is challenging, as the environments in which they act have large state spaces, and their decision-making can be affected by delayed rewards, making it difficult to analyze their behavior. To address this problem, several approaches have been developed. Some approaches attempt to convey the global behavior of the agent, describing the actions it takes in different states. Other approaches devised local explanations which provide information regarding the agent's decision-making in a particular state. In this paper, we combine global and local explanation methods, and evaluate their joint and separate contributions, providing (to the best of our knowledge) the first user study of combined local and global explanations for RL agents. Specifically, we augment strategy summaries that extract important trajectories of states from simulations of the agent with saliency maps which show what information the agent attends to. Our results show that the choice of what states to include in the summary (global information) strongly affects people's understanding of agents: participants shown summaries that included important states significantly outperformed participants who were presented with agent behavior in a randomly set of chosen world-states. We find mixed results with respect to augmenting demonstrations with saliency maps (local information), as the addition of saliency maps did not significantly improve performance in most cases. However, we do find some evidence that saliency maps can help users better understand what information the agent relies on in its decision making, suggesting avenues for future work that can further improve explanations of RL agents.

READ FULL TEXT

page 11

page 12

page 26

research
10/21/2022

Integrating Policy Summaries with Reward Decomposition for Explaining Reinforcement Learning Agents

Explaining the behavior of reinforcement learning agents operating in se...
research
03/22/2019

Explaining Reinforcement Learning to Mere Mortals: An Empirical Study

We present a user study to investigate the impact of explanations on non...
research
12/09/2019

Exploratory Not Explanatory: Counterfactual Analysis of Saliency Maps for Deep RL

Saliency maps have been used to support explanations of deep reinforceme...
research
05/30/2019

Exploring Computational User Models for Agent Policy Summarization

AI agents are being developed to support high stakes decision-making pro...
research
09/27/2019

Counterfactual States for Atari Agents via Generative Deep Learning

Although deep reinforcement learning agents have produced impressive res...
research
02/05/2021

"I Don't Think So": Disagreement-Based Policy Summaries for Comparing Agents

With Artificial Intelligence on the rise, human interaction with autonom...
research
05/06/2023

Explaining RL Decisions with Trajectories

Explanation is a key component for the adoption of reinforcement learnin...

Please sign up or login with your details

Forgot password? Click here to reset