On the Calibration of Pre-trained Language Models using Mixup Guided by Area Under the Margin and Saliency

03/14/2022
by   Seo Yeon Park, et al.
0

A well-calibrated neural model produces confidence (probability outputs) closely approximated by the expected accuracy. While prior studies have shown that mixup training as a data augmentation technique can improve model calibration on image classification tasks, little is known about using mixup for model calibration on natural language understanding (NLU) tasks. In this paper, we explore mixup for model calibration on several NLU tasks and propose a novel mixup strategy for pre-trained language models that improves model calibration further. Our proposed mixup is guided by both the Area Under the Margin (AUM) statistic (Pleiss et al., 2020) and the saliency map of each sample (Simonyan et al.,2013). Moreover, we combine our mixup strategy with model miscalibration correction techniques (i.e., label smoothing and temperature scaling) and provide detailed analyses of their impact on our proposed mixup. We focus on systematically designing experiments on three NLU tasks: natural language inference, paraphrase detection, and commonsense reasoning. Our method achieves the lowest expected calibration error compared to strong baselines on both in-domain and out-of-domain test samples while maintaining competitive accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/17/2020

Calibration of Pre-trained Transformers

Pre-trained Transformers are now ubiquitous in natural language processi...
research
05/06/2022

A Data Cartography based MixUp for Pre-trained Language Models

MixUp is a data augmentation strategy where additional samples are gener...
research
02/13/2023

Bag of Tricks for In-Distribution Calibration of Pretrained Transformers

While pre-trained language models (PLMs) have become a de-facto standard...
research
06/28/2020

A Confidence-Calibrated MOBA Game Winner Predictor

In this paper, we propose a confidence-calibration method for predicting...
research
11/06/2022

Calibration Meets Explanation: A Simple and Effective Approach for Model Confidence Estimates

Calibration strengthens the trustworthiness of black-box models by produ...
research
01/28/2022

Calibrating Histopathology Image Classifiers using Label Smoothing

The classification of histopathology images fundamentally differs from t...
research
10/31/2021

PnPOOD : Out-Of-Distribution Detection for Text Classification via Plug andPlay Data Augmentation

While Out-of-distribution (OOD) detection has been well explored in comp...

Please sign up or login with your details

Forgot password? Click here to reset