Learning Explanations from Language Data

08/13/2018
by   David Harbecke, et al.
0

PatternAttribution is a recent method, introduced in the vision domain, that explains classifications of deep neural networks. We demonstrate that it also generates meaningful interpretations in the language domain.

READ FULL TEXT

page 1

page 2

page 3

research
01/07/2022

Decision problem of some bundled FOML fragments

Over increasing domain interpretations, ∃and ∀bundled fragments are deci...
research
03/27/2021

Explaining the Road Not Taken

It is unclear if existing interpretations of deep neural network models ...
research
10/04/2020

Explaining Deep Neural Networks

Deep neural networks are becoming more and more popular due to their rev...
research
07/24/2020

Memory networks for consumer protection:unfairness exposed

Recent work has demonstrated how data-driven AI methods can leverage con...
research
04/28/2023

Interpreting Vision and Language Generative Models with Semantic Visual Priors

When applied to Image-to-text models, interpretability methods often pro...
research
07/01/2020

In-Distribution Interpretability for Challenging Modalities

It is widely recognized that the predictions of deep neural networks are...
research
05/11/2021

Leveraging Sparse Linear Layers for Debuggable Deep Networks

We show how fitting sparse linear models over learned deep feature repre...

Please sign up or login with your details

Forgot password? Click here to reset