Can Language Models Be Specific? How?

10/11/2022
by   Jie Huang, et al.
0

A good speaker not only needs to be correct, but also has the ability to be specific when desired, and so are language models. In this paper, we propose to measure how specific the language of pre-trained language models (PLMs) is. To achieve this, we introduce a novel approach to build a benchmark for specificity testing by forming masked token prediction tasks with prompts. For instance, given “J. K. Rowling was born in [MASK].”, we want to test whether a more specific answer will be better filled in by PLMs, e.g., Yate instead of England. From our evaluations, we show that existing PLMs have only a slight preference for more specific answers. We identify underlying factors affecting the specificity and design two prompt-based methods to improve the specificity. Results show that the specificity of the models can be improved by the proposed methods without additional training. We believe this work can provide new insights for language modeling and encourage the research community to further explore this important but understudied problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/28/2023

On Pre-trained Language Models for Antibody

Antibodies are vital proteins offering robust protection for the human b...
research
12/14/2021

Deciphering antibody affinity maturation with language models and weakly supervised learning

In response to pathogens, the adaptive immune system generates specific ...
research
06/01/2023

ReviewerGPT? An Exploratory Study on Using Large Language Models for Paper Reviewing

Given the rapid ascent of large language models (LLMs), we study the que...
research
09/16/2022

Negation, Coordination, and Quantifiers in Contextualized Language Models

With the success of contextualized language models, much research explor...
research
10/09/2022

Improve Transformer Pre-Training with Decoupled Directional Relative Position Encoding and Representation Differentiations

In this work, we revisit the Transformer-based pre-trained language mode...
research
04/06/2020

"You are grounded!": Latent Name Artifacts in Pre-trained Language Models

Pre-trained language models (LMs) may perpetuate biases originating in t...
research
06/16/2023

Structured Thoughts Automaton: First Formalized Execution Model for Auto-Regressive Language Models

In recent months, Language Models (LMs) have become a part of daily disc...

Please sign up or login with your details

Forgot password? Click here to reset