Mitigating belief projection in explainable artificial intelligence via Bayesian Teaching

02/07/2021
by   Scott Cheng-Hsin Yang, et al.
29

State-of-the-art deep-learning systems use decision rules that are challenging for humans to model. Explainable AI (XAI) attempts to improve human understanding but rarely accounts for how people typically reason about unfamiliar agents. We propose explicitly modeling the human explainee via Bayesian Teaching, which evaluates explanations by how much they shift explainees' inferences toward a desired goal. We assess Bayesian Teaching in a binary image classification task across a variety of contexts. Absent intervention, participants predict that the AI's classifications will match their own, but explanations generated by Bayesian Teaching improve their ability to predict the AI's judgements by moving them away from this prior belief. Bayesian Teaching further allows each case to be broken down into sub-examples (here saliency maps). These sub-examples complement whole examples by improving error detection for familiar categories, whereas whole examples help predict correct AI judgements of unfamiliar cases.

READ FULL TEXT

page 1

page 11

page 22

page 24

page 25

page 26

page 27

page 33

research
06/08/2021

Explainable AI for medical imaging: Explaining pneumothorax diagnoses with Bayesian Teaching

Limited expert time is a key bottleneck in medical imaging. Due to advan...
research
06/16/2021

Explainable AI for Natural Adversarial Images

Adversarial images highlight how vulnerable modern image classifiers are...
research
05/16/2021

Abstraction, Validation, and Generalization for Explainable Artificial Intelligence

Neural network architectures are achieving superhuman performance on an ...
research
06/21/2021

A Turing Test for Transparency

A central goal of explainable artificial intelligence (XAI) is to improv...
research
11/12/2018

TED: Teaching AI to Explain its Decisions

Artificial intelligence systems are being increasingly deployed due to t...
research
11/22/2021

Teaching Humans When To Defer to a Classifier via Exemplars

Expert decision makers are starting to rely on data-driven automated age...

Please sign up or login with your details

Forgot password? Click here to reset