How poor is the stimulus? Evaluating hierarchical generalization in neural networks trained on child-directed speech

01/26/2023
by   Aditya Yedetore, et al.
0

When acquiring syntax, children consistently choose hierarchical rules over competing non-hierarchical possibilities. Is this preference due to a learning bias for hierarchical structure, or due to more general biases that interact with hierarchical cues in children's linguistic input? We explore these possibilities by training LSTMs and Transformers - two types of neural networks without a hierarchical bias - on data similar in quantity and content to children's linguistic input: text from the CHILDES corpus. We then evaluate what these models have learned about English yes/no questions, a phenomenon for which hierarchical structure is crucial. We find that, though they perform well at capturing the surface statistics of child-directed speech (as measured by perplexity), both model types generalize in a way more consistent with an incorrect linear rule than the correct hierarchical rule. These results suggest that human-like generalization from text alone requires stronger biases than the general sequence-processing biases of standard neural network architectures.

READ FULL TEXT
research
02/25/2018

Revisiting the poverty of the stimulus: hierarchical generalization without a hierarchical bias in recurrent neural networks

Syntactic rules in human language usually refer to the hierarchical stru...
research
01/10/2020

Does syntax need to grow on trees? Sources of hierarchical inductive bias in sequence-to-sequence networks

Learners that are exposed to the same training data might generalize dif...
research
05/06/2020

Learning to Understand Child-directed and Adult-directed Speech

Speech directed to children differs from adult-directed speech in lingui...
research
02/02/2018

Order matters: Distributional properties of speech to young children bootstraps learning of semantic representations

Some researchers claim that language acquisition is critically dependent...
research
11/27/2016

The polysemy of the words that children learn over time

Here we study polysemy as a potential learning bias in vocabulary learni...
research
06/24/2019

Mutual exclusivity as a challenge for neural networks

Strong inductive biases allow children to learn in fast and adaptable wa...
research
07/14/2020

Can neural networks acquire a structural bias from raw linguistic data?

We evaluate whether BERT, a widely used neural network for sentence proc...

Please sign up or login with your details

Forgot password? Click here to reset