Testing Causal Models of Word Meaning in GPT-3 and -4

05/24/2023
by   Sam Musker, et al.
0

Large Language Models (LLMs) have driven extraordinary improvements in NLP. However, it is unclear how such models represent lexical concepts-i.e., the meanings of the words they use. This paper evaluates the lexical representations of GPT-3 and GPT-4 through the lens of HIPE theory, a theory of concept representations which focuses on representations of words describing artifacts (such as "mop", "pencil", and "whistle"). The theory posits a causal graph that relates the meanings of such words to the form, use, and history of the objects to which they refer. We test LLMs using the same stimuli originally used by Chaigneau et al. (2004) to evaluate the theory in humans, and consider a variety of prompt designs. Our experiments concern judgements about causal outcomes, object function, and object naming. We find no evidence that GPT-3 encodes the causal structure hypothesized by HIPE, but do find evidence that GPT-4 encodes such structure. The results contribute to a growing body of research characterizing the representational capacity of large language models.

READ FULL TEXT

page 4

page 20

page 21

research
11/09/2022

Collateral facilitation in humans and language models

Are the predictions of humans and language models affected by similar th...
research
03/07/2023

Can large language models build causal graphs?

Building causal graphs can be a laborious process. To ensure all relevan...
research
10/18/2022

Systematicity in GPT-3's Interpretation of Novel English Noun Compounds

Levin et al. (2019) show experimentally that the interpretations of nove...
research
07/27/2023

A Geometric Notion of Causal Probing

Large language models rely on real-valued representations of text to mak...
research
05/29/2020

A Comparative Study of Lexical Substitution Approaches based on Neural Language Models

Lexical substitution in context is an extremely powerful technology that...
research
05/28/2021

What if This Modified That? Syntactic Interventions via Counterfactual Embeddings

Neural language models exhibit impressive performance on a variety of ta...

Please sign up or login with your details

Forgot password? Click here to reset