Bag of Tricks for In-Distribution Calibration of Pretrained Transformers

02/13/2023
by   Jaeyoung Kim, et al.
0

While pre-trained language models (PLMs) have become a de-facto standard promoting the accuracy of text classification tasks, recent studies find that PLMs often predict over-confidently. Although various calibration methods have been proposed, such as ensemble learning and data augmentation, most of the methods have been verified in computer vision benchmarks rather than in PLM-based text classification tasks. In this paper, we present an empirical study on confidence calibration for PLMs, addressing three categories, including confidence penalty losses, data augmentations, and ensemble methods. We find that the ensemble model overfitted to the training set shows sub-par calibration performance and also observe that PLMs trained with confidence penalty loss have a trade-off between calibration and accuracy. Building on these observations, we propose the Calibrated PLM (CALL), a combination of calibration techniques. The CALL complements the drawbacks that may occur when utilizing a calibration method individually and boosts both classification and calibration accuracy. Design choices in CALL's training procedures are extensively studied, and we provide a detailed analysis of how calibration techniques affect the calibration performance of PLMs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/14/2022

On the Calibration of Pre-trained Language Models using Mixup Guided by Area Under the Margin and Saliency

A well-calibrated neural model produces confidence (probability outputs)...
research
07/13/2023

Unsupervised Calibration through Prior Adaptation for Text Classification using Large Language Models

A wide variety of natural language tasks are currently being addressed w...
research
06/28/2020

A Confidence-Calibrated MOBA Game Winner Predictor

In this paper, we propose a confidence-calibration method for predicting...
research
01/14/2022

Model Stability with Continuous Data Updates

In this paper, we study the "stability" of machine learning (ML) models ...
research
10/24/2020

PEP: Parameter Ensembling by Perturbation

Ensembling is now recognized as an effective approach for increasing the...
research
10/31/2022

A Close Look into the Calibration of Pre-trained Language Models

Pre-trained language models (PLMs) achieve remarkable performance on man...
research
05/22/2023

Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot Text Classification Tasks

Text classification tasks often encounter few shot scenarios with limite...

Please sign up or login with your details

Forgot password? Click here to reset