IFAN: An Explainability-Focused Interaction Framework for Humans and NLP Models

03/06/2023
by   Edoardo Mosca, et al.
0

Interpretability and human oversight are fundamental pillars of deploying complex NLP models into real-world applications. However, applying explainability and human-in-the-loop methods requires technical proficiency. Despite existing toolkits for model understanding and analysis, options to integrate human feedback are still limited. We propose IFAN, a framework for real-time explanation-based interaction with NLP models. Through IFAN's interface, users can provide feedback to selected model explanations, which is then integrated through adapter layers to align the model with human rationale. We show the system to be effective in debiasing a hate speech classifier with minimal performance loss. IFAN also offers a visual admin system and API to manage models (and datasets) as well as control access rights. A demo is live at https://ifan.ml/

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/30/2022

XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models

NLP models are susceptible to learning spurious biases (i.e., bugs) that...
research
03/06/2021

Putting Humans in the Natural Language Processing Loop: A Survey

How can we design Natural Language Processing (NLP) systems that learn f...
research
06/29/2021

On the Interaction of Belief Bias and Explanations

A myriad of explainability methods have been proposed in recent years, b...
research
05/10/2022

Sensible AI: Re-imagining Interpretability and Explainability using Sensemaking Theory

Understanding how ML models work is a prerequisite for responsibly desig...
research
09/07/2023

Automatic Concept Embedding Model (ACEM): No train-time concepts, No issue!

Interpretability and explainability of neural networks is continuously i...
research
08/31/2021

Thermostat: A Large Collection of NLP Model Explanations and Analysis Tools

In the language domain, as in other domains, neural explainability takes...
research
07/01/2020

Unifying Model Explainability and Robustness via Machine-Checkable Concepts

As deep neural networks (DNNs) get adopted in an ever-increasing number ...

Please sign up or login with your details

Forgot password? Click here to reset