Exploring BERT's Sensitivity to Lexical Cues using Tests from Semantic Priming

10/06/2020
by   Kanishka Misra, et al.
0

Models trained to estimate word probabilities in context have become ubiquitous in natural language processing. How do these models use lexical cues in context to inform their word probabilities? To answer this question, we present a case study analyzing the pre-trained BERT model with tests informed by semantic priming. Using English lexical stimuli that show priming in humans, we find that BERT too shows "priming," predicting a word with greater probability when the context includes a related word versus an unrelated one. This effect decreases as the amount of information provided by the context increases. Follow-up analysis shows BERT to be increasingly distracted by related prime words as context becomes more informative, assigning lower probabilities to related words. Our findings highlight the importance of considering contextual constraint effects when studying word prediction in these models, and highlight possible parallels with human processing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/24/2020

MULTISEM at SemEval-2020 Task 3: Fine-tuning BERT for Lexical Meaning

We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded...
research
12/13/2021

Context vs Target Word: Quantifying Biases in Lexical Semantic Datasets

State-of-the-art contextualized models such as BERT use tasks such as Wi...
research
02/08/2022

HistBERT: A Pre-trained Language Model for Diachronic Lexical Semantic Analysis

Contextualized word embeddings have demonstrated state-of-the-art perfor...
research
10/05/2020

Speakers Fill Lexical Semantic Gaps with Context

Lexical ambiguity is widespread in language, allowing for the reuse of e...
research
05/03/2020

An Accurate Model for Predicting the (Graded) Effect of Context in Word Similarity Based on Bert

Natural Language Processing (NLP) has been widely used in the semantic a...
research
04/19/2021

BigGreen at SemEval-2021 Task 1: Lexical Complexity Prediction with Assembly Models

This paper describes a system submitted by team BigGreen to LCP 2021 for...
research
02/14/2023

Exploring Category Structure with Contextual Language Models and Lexical Semantic Networks

Recent work on predicting category structure with distributional models,...

Please sign up or login with your details

Forgot password? Click here to reset