Physics-Inspired Interpretability Of Machine Learning Models

04/05/2023
by   Maximilian P Niroomand, et al.
0

The ability to explain decisions made by machine learning models remains one of the most significant hurdles towards widespread adoption of AI in highly sensitive areas such as medicine, cybersecurity or autonomous driving. Great interest exists in understanding which features of the input data prompt model decision making. In this contribution, we propose a novel approach to identify relevant features of the input data, inspired by methods from the energy landscapes field, developed in the physical sciences. By identifying conserved weights within groups of minima of the loss landscapes, we can identify the drivers of model decision making. Analogues to this idea exist in the molecular sciences, where coordinate invariants or order parameters are employed to identify critical features of a molecule. However, no such approach exists for machine learning loss landscapes. We will demonstrate the applicability of energy landscape methods to machine learning models and give examples, both synthetic and from the real world, for how these methods can help to make models more interpretable.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2017

Perspective: Energy Landscapes for Machine Learning

Machine learning techniques are being increasingly used as flexible non-...
research
06/21/2023

Investigating Poor Performance Regions of Black Boxes: LIME-based Exploration in Sepsis Detection

Interpreting machine learning models remains a challenge, hindering thei...
research
11/20/2018

A Gray Box Interpretable Visual Debugging Approach for Deep Sequence Learning Model

Deep Learning algorithms are often used as black box type learning and t...
research
07/12/2022

Revealing Unfair Models by Mining Interpretable Evidence

The popularity of machine learning has increased the risk of unfair mode...
research
09/18/2020

Probably Approximately Correct Explanations of Machine Learning Models via Syntax-Guided Synthesis

We propose a novel approach to understanding the decision making of comp...
research
07/10/2018

Fooling the classifier: Ligand antagonism and adversarial examples

Machine learning algorithms are sensitive to so-called adversarial pertu...

Please sign up or login with your details

Forgot password? Click here to reset