Lower Perplexity is Not Always Human-Like

06/02/2021
by   Tatsuki Kuribayashi, et al.
0

In computational psycholinguistics, various language models have been evaluated against human reading behavior (e.g., eye movement) to build human-like computational models. However, most previous efforts have focused almost exclusively on English, despite the recent trend towards linguistic universal within the general community. In order to fill the gap, this paper investigates whether the established results in computational psycholinguistics can be generalized across languages. Specifically, we re-examine an established generalization – the lower perplexity a language model has, the more human-like the language model is – in Japanese with typologically different structures from English. Our experiments demonstrate that this established generalization exhibits a surprising lack of universality; namely, lower perplexity is not always human-like. Moreover, this discrepancy between English and Japanese is further explored from the perspective of (non-)uniform information density. Overall, our results suggest that a cross-lingual evaluation will be necessary to construct human-like computational models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/23/2023

Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning

Most Transformer language models are primarily pretrained on English tex...
research
06/02/2021

Uncovering Constraint-Based Behavior in Neural Models via Targeted Fine-Tuning

A growing body of literature has focused on detailing the linguistic kno...
research
03/16/2022

Cross-Lingual Ability of Multilingual Masked Language Models: A Study of Language Structure

Multilingual pre-trained language models, such as mBERT and XLM-R, have ...
research
11/14/2022

Speaking Multiple Languages Affects the Moral Bias of Language Models

Pre-trained multilingual language models (PMLMs) are commonly used when ...
research
06/05/2023

Second Language Acquisition of Neural Language Models

With the success of neural language models (LMs), their language acquisi...
research
06/12/2023

Large language models and (non-)linguistic recursion

Recursion is one of the hallmarks of human language. While many design f...
research
09/23/2021

Revisiting the Uniform Information Density Hypothesis

The uniform information density (UID) hypothesis posits a preference amo...

Please sign up or login with your details

Forgot password? Click here to reset