Hierarchical Symbolic Reasoning in Hyperbolic Space for Deep Discriminative Models

Explanations for black-box models help us understand model decisions as well as provide information on model biases and inconsistencies. Most of the current explainability techniques provide a single level of explanation, often in terms of feature importance scores or feature attention maps in input space. Our focus is on explaining deep discriminative models at multiple levels of abstraction, from fine-grained to fully abstract explanations. We achieve this by using the natural properties of hyperbolic geometry to more efficiently model a hierarchy of symbolic features and generate hierarchical symbolic rules as part of our explanations. Specifically, for any given deep discriminative model, we distill the underpinning knowledge by discretisation of the continuous latent space using vector quantisation to form symbols, followed by a hyperbolic reasoning block to induce an abstraction tree. We traverse the tree to extract explanations in terms of symbolic rules and its corresponding visual semantics. We demonstrate the effectiveness of our method on the MNIST and AFHQ high-resolution animal faces dataset. Our framework is available at <https://github.com/koriavinash1/SymbolicInterpretability>.

READ FULL TEXT

page 8

page 9

page 16

page 17

page 18

research
07/05/2022

GLANCE: Global to Local Architecture-Neutral Concept-based Explanations

Most of the current explainability techniques focus on capturing the imp...
research
10/17/2022

Visual Debates

The natural way of obtaining different perspectives on any given topic i...
research
05/16/2021

Expressive Explanations of DNNs by Combining Concept Analysis with ILP

Explainable AI has emerged to be a key component for black-box machine l...
research
06/23/2022

ASTERYX : A model-Agnostic SaT-basEd appRoach for sYmbolic and score-based eXplanations

The ever increasing complexity of machine learning techniques used more ...
research
07/28/2023

SAFE: Saliency-Aware Counterfactual Explanations for DNN-based Automated Driving Systems

A CF explainer identifies the minimum modifications in the input that wo...
research
04/11/2022

medXGAN: Visual Explanations for Medical Classifiers through a Generative Latent Space

Despite the surge of deep learning in the past decade, some users are sk...
research
12/23/2021

Prolog-based agnostic explanation module for structured pattern classification

This paper presents a Prolog-based reasoning module to generate counterf...

Please sign up or login with your details

Forgot password? Click here to reset