CUE: An Uncertainty Interpretation Framework for Text Classifiers Built on Pre-Trained Language Models

06/06/2023
by   Jiazheng Li, et al.
0

Text classifiers built on Pre-trained Language Models (PLMs) have achieved remarkable progress in various tasks including sentiment analysis, natural language inference, and question-answering. However, the occurrence of uncertain predictions by these classifiers poses a challenge to their reliability when deployed in practical applications. Much effort has been devoted to designing various probes in order to understand what PLMs capture. But few studies have delved into factors influencing PLM-based classifiers' predictive uncertainty. In this paper, we propose a novel framework, called CUE, which aims to interpret uncertainties inherent in the predictions of PLM-based models. In particular, we first map PLM-encoded representations to a latent space via a variational auto-encoder. We then generate text representations by perturbing the latent space which causes fluctuation in predictive uncertainty. By comparing the difference in predictive uncertainty between the perturbed and the original text representations, we are able to identify the latent dimensions responsible for uncertainty and subsequently trace back to the input features that contribute to such uncertainty. Our extensive experiments on four benchmark datasets encompassing linguistic acceptability classification, emotion classification, and natural language inference show the feasibility of our proposed framework. Our source code is available at: https://github.com/lijiazheng99/CUE.

READ FULL TEXT

page 2

page 18

research
01/11/2022

Explaining Prediction Uncertainty of Pre-trained Language Models by Detecting Uncertain Words in Inputs

Estimating the predictive uncertainty of pre-trained language models is ...
research
07/01/2022

Is neural language acquisition similar to natural? A chronological probing study

The probing methodology allows one to obtain a partial representation of...
research
12/17/2020

MASKER: Masked Keyword Regularization for Reliable Text Classification

Pre-trained language models have achieved state-of-the-art accuracies on...
research
03/27/2021

You Can Do Better! If You Elaborate the Reason When Making Prediction

Neural predictive models have achieved groundbreaking performance improv...
research
02/19/2023

Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation

We introduce a method to measure uncertainty in large language models. F...
research
06/01/2023

Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction

Recently, aspect sentiment quad prediction has received widespread atten...
research
05/03/2023

Robust Natural Language Watermarking through Invariant Features

Recent years have witnessed a proliferation of valuable original natural...

Please sign up or login with your details

Forgot password? Click here to reset