Calibration, Entropy Rates, and Memory in Language Models

06/11/2019
by   Mark Braverman, et al.
1

Building accurate language models that capture meaningful long-term dependencies is a core challenge in natural language processing. Towards this end, we present a calibration-based approach to measure long-term discrepancies between a generative sequence model and the true distribution, and use these discrepancies to improve the model. Empirically, we show that state-of-the-art language models, including LSTMs and Transformers, are miscalibrated: the entropy rates of their generations drift dramatically upward over time. We then provide provable methods to mitigate this phenomenon. Furthermore, we show how this calibration-based approach can also be used to measure the amount of memory that language models use for prediction.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2023

A Comparative Study of Pretrained Language Models for Long Clinical Text

Objective: Clinical knowledge enriched transformer models (e.g., Clinica...
research
05/04/2023

Conformal Nucleus Sampling

Language models generate text based on successively sampling the next wo...
research
05/19/2023

Extending Memory for Language Modelling

Breakthroughs in deep learning and memory networks have made major advan...
research
05/23/2023

Narrative XL: A Large-scale Dataset For Long-Term Memory Models

Despite their tremendous successes, most large language models do not ha...
research
01/26/2022

An Assessment of the Impact of OCR Noise on Language Models

Neural language models are the backbone of modern-day natural language p...
research
05/29/2023

Marked Personas: Using Natural Language Prompts to Measure Stereotypes in Language Models

To recognize and mitigate harms from large language models (LLMs), we ne...
research
10/07/2022

Novice Type Error Diagnosis with Natural Language Models

Strong static type systems help programmers eliminate many errors withou...

Please sign up or login with your details

Forgot password? Click here to reset