Inflecting when there's no majority: Limitations of encoder-decoder neural networks as cognitive models for German plurals

05/18/2020
by   Kate McCurdy, et al.
0

Can artificial neural networks learn to represent inflectional morphology and generalize to new words as human speakers do? Kirov and Cotterell (2018) argue that the answer is yes: modern Encoder-Decoder (ED) architectures learn human-like behavior when inflecting English verbs, such as extending the regular past tense form -(e)d to novel words. However, their work does not address the criticism raised by Marcus et al. (1995): that neural models may learn to extend not the regular, but the most frequent class – and thus fail on tasks like German number inflection, where infrequent suffixes like -s can still be productively generalized. To investigate this question, we first collect a new dataset from German speakers (production and ratings of plural forms for novel nouns) that is designed to avoid sources of information unavailable to the ED model. The speaker data show high variability, and two suffixes evince 'regular' behavior, appearing more often with phonologically atypical inputs. Encoder-decoder models do generalize the most frequently produced plural class, but do not show human-like variability or 'regular' extension of these other plural markers. We conclude that modern neural models may still struggle with minority-class generalization.

READ FULL TEXT

page 7

page 8

research
10/22/2022

A Comprehensive Comparison of Neural Networks as Cognitive Models of Inflection

Neural networks have long been at the center of a debate around the cogn...
research
06/04/2019

Are we there yet? Encoder-decoder neural networks as cognitive models of English past tense inflection

The cognitive mechanisms needed to account for the English past tense ha...
research
04/25/2022

ED2LM: Encoder-Decoder to Language Model for Faster Document Re-ranking Inference

State-of-the-art neural models typically encode document-query pairs usi...
research
07/12/2018

Recurrent Neural Networks in Linguistic Theory: Revisiting Pinker and Prince (1988) and the Past Tense Debate

Can advances in NLP help advance cognitive modeling? We examine the role...
research
10/17/2022

How do we get there? Evaluating transformer neural networks as cognitive models for English past tense inflection

There is an ongoing debate on whether neural networks can grasp the quas...
research
08/09/2021

Not quite there yet: Combining analogical patterns and encoder-decoder networks for cognitively plausible inflection

The paper presents four models submitted to Part 2 of the SIGMORPHON 202...
research
09/18/2020

Forecasting time series with encoder-decoder neural networks

In this paper, we consider high-dimensional stationary processes where a...

Please sign up or login with your details

Forgot password? Click here to reset