Teaching Models to Express Their Uncertainty in Words

05/28/2022
by   Stephanie Lin, et al.
0

We show that a GPT-3 model can learn to express uncertainty about its own answers in natural language – without use of model logits. When given a question, the model generates both an answer and a level of confidence (e.g. "90 are well calibrated. The model also remains moderately calibrated under distribution shift, and is sensitive to uncertainty in its own answers, rather than imitating human examples. To our knowledge, this is the first time a model has been shown to express calibrated uncertainty about its own answers in natural language. For testing calibration, we introduce the CalibratedMath suite of tasks. We compare the calibration of uncertainty expressed in words ("verbalized probability") to uncertainty extracted from model logits. Both kinds of uncertainty are capable of generalizing calibration under distribution shift. We also provide evidence that GPT-3's ability to generalize calibration depends on pre-trained latent representations that correlate with epistemic uncertainty over its answers.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/17/2020

Calibration of Pre-trained Transformers

Pre-trained Transformers are now ubiquitous in natural language processi...
research
09/21/2021

Bayesian Confidence Calibration for Epistemic Uncertainty Modelling

Modern neural networks have found to be miscalibrated in terms of confid...
research
12/14/2020

Improving model calibration with accuracy versus uncertainty optimization

Obtaining reliable and accurate quantification of uncertainty estimates ...
research
02/26/2023

Navigating the Grey Area: Expressions of Overconfidence and Uncertainty in Language Models

Despite increasingly fluent, relevant, and coherent language generation,...
research
12/01/2022

Deep Kernel Learning for Mortality Prediction in the Face of Temporal Shift

Neural models, with their ability to provide novel representations, have...
research
05/19/2023

What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability

In Natural Language Generation (NLG) tasks, for any input, multiple comm...
research
10/31/2022

A Close Look into the Calibration of Pre-trained Language Models

Pre-trained language models (PLMs) achieve remarkable performance on man...

Please sign up or login with your details

Forgot password? Click here to reset