TripleTree: A Versatile Interpretable Representation of Black Box Agents and their Environments

09/10/2020
by   Tom Bewley, et al.
15

In explainable artificial intelligence, there is increasing interest in understanding the behaviour of autonomous agents to build trust and validate performance. Modern agent architectures, such as those trained by deep reinforcement learning, are currently so lacking in interpretable structure as to effectively be black boxes, but insights may still be gained from an external, behaviourist perspective. Inspired by conceptual spaces theory, we suggest that a versatile first step towards general understanding is to discretise the state space into convex regions, jointly capturing similarities over the agent's action, value function and temporal dynamics within a dataset of observations. We create such a representation using a novel variant of the CART decision tree algorithm, and demonstrate how it facilitates practical understanding of black box agents through prediction, visualisation and rule-based explanation.

READ FULL TEXT

page 5

page 6

page 7

research
07/02/2020

Am I Building a White Box Agent or Interpreting a Black Box Agent?

The rule extraction literature contains the notion of a fidelity-accurac...
research
09/20/2020

Interpretable-AI Policies using Evolutionary Nonlinear Decision Trees for Discrete Action Systems

Black-box artificial intelligence (AI) induction methods such as deep re...
research
06/07/2021

Explainable Artificial Intelligence (XAI) for Increasing User Trust in Deep Reinforcement Learning Driven Autonomous Systems

We consider the problem of providing users of deep Reinforcement Learnin...
research
09/07/2023

Learning of Generalizable and Interpretable Knowledge in Grid-Based Reinforcement Learning Environments

Understanding the interactions of agents trained with deep reinforcement...
research
11/06/2022

ProtoX: Explaining a Reinforcement Learning Agent via Prototyping

While deep reinforcement learning has proven to be successful in solving...
research
10/22/2021

ProtoShotXAI: Using Prototypical Few-Shot Architecture for Explainable AI

Unexplainable black-box models create scenarios where anomalies cause de...
research
03/01/2022

Explaining a Deep Reinforcement Learning Docking Agent Using Linear Model Trees with User Adapted Visualization

Deep neural networks (DNNs) can be useful within the marine robotics fie...

Please sign up or login with your details

Forgot password? Click here to reset