Predicting Human Psychometric Properties Using Computational Language Models

05/12/2022
by   Antonio Laverghetta Jr., et al.
9

Transformer-based language models (LMs) continue to achieve state-of-the-art performance on natural language processing (NLP) benchmarks, including tasks designed to mimic human-inspired "commonsense" competencies. To better understand the degree to which LMs can be said to have certain linguistic reasoning skills, researchers are beginning to adapt the tools and concepts from psychometrics. But to what extent can benefits flow in the other direction? In other words, can LMs be of use in predicting the psychometric properties of test items, when those items are given to human participants? If so, the benefit for psychometric practitioners is enormous, as it can reduce the need for multiple rounds of empirical testing. We gather responses from numerous human participants and LMs (transformer- and non-transformer-based) on a broad diagnostic test of linguistic competencies. We then use the human responses to calculate standard psychometric properties of the items in the diagnostic test, using the human responses and the LM responses separately. We then determine how well these two sets of predictions correlate. We find that transformer-based LMs predict the human psychometric data consistently well across most categories, suggesting that they can be used to gather human-like psychometric data without the need for extensive human trials.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2021

Can Transformer Language Models Predict Psychometric Properties?

Transformer-based language models (LMs) continue to advance state-of-the...
research
04/29/2022

Developmental Negation Processing in Transformer Language Models

Reasoning using negation is known to be difficult for transformer-based ...
research
11/18/2020

Do Fine-tuned Commonsense Language Models Really Generalize?

Recently, transformer-based methods such as RoBERTa and GPT-3 have led t...
research
04/10/2020

On the Existence of Tacit Assumptions in Contextualized Language Models

Humans carry stereotypic tacit assumptions (STAs) (Prince, 1978), or pro...
research
05/19/2023

Scaling laws for language encoding models in fMRI

Representations from transformer-based unidirectional language models ar...
research
07/07/2022

Neural Language Models are not Born Equal to Fit Brain Data, but Training Helps

Neural Language Models (NLMs) have made tremendous advances during the l...
research
05/06/2021

Do language models learn typicality judgments from text?

Building on research arguing for the possibility of conceptual and categ...

Please sign up or login with your details

Forgot password? Click here to reset