Language Models Can Learn Exceptions to Syntactic Rules

06/09/2023
by   Cara Su-Yi Leong, et al.
0

Artificial neural networks can generalize productively to novel contexts. Can they also learn exceptions to those productive rules? We explore this question using the case of restrictions on English passivization (e.g., the fact that "The vacation lasted five days" is grammatical, but "*Five days was lasted by the vacation" is not). We collect human acceptability judgments for passive sentences with a range of verbs, and show that the probability distribution defined by GPT-2, a language model, matches the human judgments with high correlation. We also show that the relative acceptability of a verb in the active vs. passive voice is positively correlated with the relative frequency of its occurrence in those voices. These results provide preliminary support for the entrenchment hypothesis, according to which learners track and uses the distributional properties of their input to learn negative exceptions to rules. At the same time, this hypothesis fails to explain the magnitude of unpassivizability demonstrated by certain individual verbs, suggesting that other cues to exceptionality are available in the linguistic input.

READ FULL TEXT
research
05/01/2020

Recurrent Neural Network Language Models Always Learn English-Like Relative Clause Attachment

A standard approach to evaluating language models analyzes how models as...
research
10/21/2022

Syntactic Surprisal From Neural Models Predicts, But Underestimates, Human Processing Difficulty From Syntactic Ambiguities

Humans exhibit garden path effects: When reading sentences that are temp...
research
05/25/2023

Passive learning of active causal strategies in agents and language models

What can be learned about causality and experimentation from passive dat...
research
03/29/2018

Colorless green recurrent networks dream hierarchically

Recurrent neural networks (RNNs) have achieved impressive results in a v...
research
12/02/2022

Event knowledge in large language models: the gap between the impossible and the unlikely

People constantly use language to learn about the world. Computational l...
research
02/25/2018

Revisiting the poverty of the stimulus: hierarchical generalization without a hierarchical bias in recurrent neural networks

Syntactic rules in human language usually refer to the hierarchical stru...
research
10/03/2015

P-trac Procedure: The Dispersion and Neutralization of Contrasts in Lexicon

Cognitive acoustic cues have an important role in shaping the phonologic...

Please sign up or login with your details

Forgot password? Click here to reset