Counterfactual Explanation and Causal Inference in Service of Robustness in Robot Control

09/18/2020
by   Simón C. Smith, et al.
14

We propose an architecture for training generative models of counterfactual conditionals of the form, 'can we modify event A to cause B instead of C?', motivated by applications in robot control. Using an 'adversarial training' paradigm, an image-based deep neural network model is trained to produce small and realistic modifications to an original image in order to cause user-defined effects. These modifications can be used in the design process of image-based robust control - to determine the ability of the controller to return to a working regime by modifications in the input space, rather than by adaptation. In contrast to conventional control design approaches, where robustness is quantified in terms of the ability to reject noise, we explore the space of counterfactuals that might cause a certain requirement to be violated, thus proposing an alternative model that might be more expressive in certain robotics applications. So, we propose the generation of counterfactuals as an approach to explanation of black-box models and the envisioning of potential movement paths in autonomous robotic control. Firstly, we demonstrate this approach in a set of classification tasks, using the well known MNIST and CelebFaces Attributes datasets. Then, addressing multi-dimensional regression, we demonstrate our approach in a reaching task with a physical robot, and in a navigation task with a robot in a digital twin simulation.

READ FULL TEXT

page 1

page 4

page 5

page 6

page 7

research
06/26/2020

Counterfactual explanation of machine learning survival models

A method for counterfactual explanation of machine learning survival mod...
research
10/28/2021

Counterfactual Explanation of Brain Activity Classifiers using Image-to-Image Transfer by Generative Adversarial Network

Deep neural networks (DNNs) can accurately decode task-related informati...
research
12/02/2019

EMAP: Explanation by Minimal Adversarial Perturbation

Modern instance-based model-agnostic explanation methods (LIME, SHAP, L2...
research
03/31/2023

Online Modifications for Event-based Signal Temporal Logic Specifications

In this paper we present a grammar and control synthesis framework for o...
research
03/02/2023

Counterfactual Edits for Generative Evaluation

Evaluation of generative models has been an underrepresented field despi...
research
06/28/2021

Contrastive Counterfactual Visual Explanations With Overdetermination

A novel explainable AI method called CLEAR Image is introduced in this p...

Please sign up or login with your details

Forgot password? Click here to reset