Explanation Augmented Feedback in Human-in-the-Loop Reinforcement Learning

07/17/2020
by   Lin Guan, et al.
0

Human-in-the-loop Reinforcement Learning (HRL) aims to integrate human guidance with Reinforcement Learning (RL) algorithms to improve sample efficiency and performance. A common type of human guidance in HRL is binary evaluative "good" or "bad" feedback for queried states and actions. However, this type of learning scheme suffers from the problems of weak supervision and poor efficiency in leveraging human feedback. To address this, we present EXPAND (EXPlanation AugmeNted feeDback) which provides a visual explanation in the form of saliency maps from humans in addition to the binary feedback. EXPAND employs a state perturbation approach based on salient information in the state to augment the binary feedback. We choose five tasks, namely Pixel-Taxi and four Atari games, to evaluate this approach. We demonstrate the effectiveness of our method using two metrics: environment sample efficiency and human feedback sample efficiency. We show that our method significantly outperforms previous methods. We also analyze the results qualitatively by visualizing the agent's attention. Finally, we present an ablation study to confirm our hypothesis that augmenting binary feedback with state salient information results in a boost in performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2018

DQN-TAMER: Human-in-the-Loop Reinforcement Learning with Intractable Feedback

Exploration has been one of the greatest challenges in reinforcement lea...
research
06/30/2020

Accelerating Reinforcement Learning Agent with EEG-based Implicit Human Feedback

Providing Reinforcement Learning (RL) agents with human feedback can dra...
research
10/07/2022

Advice Conformance Verification by Reinforcement Learning agents for Human-in-the-Loop

Human-in-the-loop (HiL) reinforcement learning is gaining traction in do...
research
06/27/2022

Humans are not Boltzmann Distributions: Challenges and Opportunities for Modelling Human Feedback and Interaction in Reinforcement Learning

Reinforcement learning (RL) commonly assumes access to well-specified re...
research
10/28/2022

When Life Gives You Lemons, Make Cherryade: Converting Feedback from Bad Responses into Good Labels

Deployed dialogue agents have the potential to integrate human feedback ...
research
08/03/2021

Accelerating the Convergence of Human-in-the-Loop Reinforcement Learning with Counterfactual Explanations

The capability to interactively learn from human feedback would enable r...
research
09/06/2021

Enhancing Visual Dialog Questioner with Entity-based Strategy Learning and Augmented Guesser

Considering the importance of building a good Visual Dialog (VD) Questio...

Please sign up or login with your details

Forgot password? Click here to reset