Exploratory Not Explanatory: Counterfactual Analysis of Saliency Maps for Deep RL

12/09/2019
by   Akanksha Atrey, et al.
20

Saliency maps have been used to support explanations of deep reinforcement learning (RL) agent behavior over temporally extended sequences. However, their use in the community indicates that the explanations derived from saliency maps are often unfalsifiable and can be highly subjective. We introduce an empirical approach grounded in counterfactual reasoning to test the hypotheses generated from saliency maps and assess the degree to which saliency maps represent semantics of RL environments. We evaluate three types of saliency maps using Atari games, a common benchmark for deep RL. Our results show the extent to which existing claims about Atari games can be evaluated and suggest that saliency maps are an exploratory tool not an explanatory tool.

READ FULL TEXT

page 2

page 7

page 14

research
12/23/2019

Explain Your Move: Understanding Agent Actions Using Focused Feature Saliency

As deep reinforcement learning (RL) is applied to more tasks, there is a...
research
01/18/2021

Benchmarking Perturbation-based Saliency Maps for Explaining Deep Reinforcement Learning Agents

Recent years saw a plethora of work on explaining complex intelligent ag...
research
03/22/2019

Explaining Reinforcement Learning to Mere Mortals: An Empirical Study

We present a user study to investigate the impact of explanations on non...
research
07/12/2019

Saliency Maps Generation for Automatic Text Summarization

Saliency map generation techniques are at the forefront of explainable A...
research
05/18/2020

Local and Global Explanations of Agent Behavior: Integrating Strategy Summaries with Saliency Maps

With advances in reinforcement learning (RL), agents are now being devel...
research
05/18/2022

Policy Distillation with Selective Input Gradient Regularization for Efficient Interpretability

Although deep Reinforcement Learning (RL) has proven successful in a wid...
research
08/07/2019

Free-Lunch Saliency via Attention in Atari Agents

We propose a new approach to visualize saliency maps for deep neural net...

Please sign up or login with your details

Forgot password? Click here to reset