Counterfactual Interventions Reveal the Causal Effect of Relative Clause Representations on Agreement Prediction

05/14/2021
by   Shauli Ravfogel, et al.
24

When language models process syntactically complex sentences, do they use abstract syntactic information present in these sentences in a manner that is consistent with the grammar of English, or do they rely solely on a set of heuristics? We propose a method to tackle this question, AlterRep. For any linguistic feature in the sentence, AlterRep allows us to generate counterfactual representations by altering how this feature is encoded, while leaving all other aspects of the original representation intact. Then, by measuring the change in a models' word prediction with these counterfactual representations in different sentences, we can draw causal conclusions about the contexts in which the model uses the linguistic feature (if any). Applying this method to study how BERT uses relative clause (RC) span information, we found that BERT uses information about RC spans during agreement prediction using the linguistically correct strategy. We also found that counterfactual representations generated for a specific RC subtype influenced the number prediction in sentences with other RC subtypes, suggesting that information about RC boundaries was encoded abstractly in BERT's representation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2021

What if This Modified That? Syntactic Interventions via Counterfactual Embeddings

Neural language models exhibit impressive performance on a variety of ta...
research
06/01/2020

When Bert Forgets How To POS: Amnesic Probing of Linguistic Properties and MLM Predictions

A growing body of work makes use of probing in order to investigate the ...
research
10/25/2022

Causal Analysis of Syntactic Agreement Neurons in Multilingual Language Models

Structural probing work has found evidence for latent syntactic informat...
research
04/19/2022

Probing for the Usage of Grammatical Number

A central quest of probing is to uncover how pre-trained models encode a...
research
09/14/2021

Frequency Effects on Syntactic Rule Learning in Transformers

Pre-trained language models perform well on a variety of linguistic task...
research
03/01/2023

Competence-Based Analysis of Language Models

Despite the recent success of large pretrained language models (LMs) on ...
research
10/21/2022

Probing with Noise: Unpicking the Warp and Weft of Embeddings

Improving our understanding of how information is encoded in vector spac...

Please sign up or login with your details

Forgot password? Click here to reset