Policy Optimization with Sparse Global Contrastive Explanations

07/13/2022
by   Jiayu Yao, et al.
0

We develop a Reinforcement Learning (RL) framework for improving an existing behavior policy via sparse, user-interpretable changes. Our goal is to make minimal changes while gaining as much benefit as possible. We define a minimal change as having a sparse, global contrastive explanation between the original and proposed policy. We improve the current policy with the constraint of keeping that global contrastive explanation short. We demonstrate our framework with a discrete MDP and a continuous 2D navigation domain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/14/2022

(When) Are Contrastive Explanations of Reinforcement Learning Helpful?

Global explanations of a reinforcement learning (RL) agent's expected be...
research
07/25/2023

Counterfactual Explanation Policies in RL

As Reinforcement Learning (RL) agents are increasingly employed in diver...
research
02/19/2020

Learning Global Transparent Models from Local Contrastive Explanations

There is a rich and growing literature on producing local point wise con...
research
05/29/2023

Reason to explain: Interactive contrastive explanations (REASONX)

Many high-performing machine learning models are not interpretable. As t...
research
07/23/2018

Contrastive Explanations for Reinforcement Learning in terms of Expected Consequences

Machine Learning models become increasingly proficient in complex tasks....
research
07/06/2021

AdaRL: What, Where, and How to Adapt in Transfer Reinforcement Learning

Most approaches in reinforcement learning (RL) are data-hungry and speci...
research
09/15/2021

On the Limits of Minimal Pairs in Contrastive Evaluation

Minimal sentence pairs are frequently used to analyze the behavior of la...

Please sign up or login with your details

Forgot password? Click here to reset