A study on the Interpretability of Neural Retrieval Models using DeepSHAP

07/15/2019
by   Zeon Trevor Fernando, et al.
0

A recent trend in IR has been the usage of neural networks to learn retrieval models for text based adhoc search. While various approaches and architectures have yielded significantly better performance than traditional retrieval models such as BM25, it is still difficult to understand exactly why a document is relevant to a query. In the ML community several approaches for explaining decisions made by deep neural networks have been proposed -- including DeepSHAP which modifies the DeepLift algorithm to estimate the relative importance (shapley values) of input features for a given decision by comparing the activations in the network for a given image against the activations caused by a reference input. In image classification, the reference input tends to be a plain black image. While DeepSHAP has been well studied for image classification tasks, it remains to be seen how we can adapt it to explain the output of Neural Retrieval Models (NRMs). In particular, what is a good "black" image in the context of IR? In this paper we explored various reference input document construction techniques. Additionally, we compared the explanations generated by DeepSHAP to LIME (a model agnostic approach) and found that the explanations differ considerably. Our study raises concerns regarding the robustness and accuracy of explanations produced for NRMs. With this paper we aim to shed light on interesting problems surrounding interpretability in NRMs and highlight areas of future work.

READ FULL TEXT
research
06/15/2021

Towards Axiomatic Explanations for Neural Ranking Models

Recently, neural networks have been successfully employed to improve upo...
research
07/20/2018

Explaining Image Classifiers by Adaptive Dropout and Generative In-filling

Explanations of black-box classifiers often rely on saliency maps, which...
research
10/17/2019

Effect of Superpixel Aggregation on Explanations in LIME – A Case Study with Biological Data

End-to-end learning with deep neural networks, such as convolutional neu...
research
07/19/2023

TbExplain: A Text-based Explanation Method for Scene Classification Models with the Statistical Prediction Correction

The field of Explainable Artificial Intelligence (XAI) aims to improve t...
research
05/27/2020

Explaining Neural Networks by Decoding Layer Activations

To derive explanations for deep learning models, ie. classifiers, we pro...
research
12/08/2015

Explaining NonLinear Classification Decisions with Deep Taylor Decomposition

Nonlinear methods such as Deep Neural Networks (DNNs) are the gold stand...
research
02/28/2021

Model-Agnostic Explainability for Visual Search

What makes two images similar? We propose new approaches to generate mod...

Please sign up or login with your details

Forgot password? Click here to reset