What can generic neural networks learn from a child's visual experience?

05/24/2023
by   A. Emin Orhan, et al.
0

Young children develop sophisticated internal models of the world based on their egocentric visual experience. How much of this is driven by innate constraints and how much is driven by their experience? To investigate these questions, we train state-of-the-art neural networks on a realistic proxy of a child's visual experience without any explicit supervision or domain-specific inductive biases. Specifically, we train both embedding models and generative models on 200 hours of headcam video from a single child collected over two years. We train a total of 72 different models, exploring a range of model architectures and self-supervised learning algorithms, and comprehensively evaluate their performance in downstream tasks. The best embedding models perform at 70 also learn broad semantic categories without any labeled examples and learn to localize semantic categories in an image without any location supervision. However, these models are less object-centric and more background-sensitive than comparable ImageNet-trained models. Generative models trained with the same data successfully extrapolate simple properties of partially masked objects, such as their texture, color, orientation, and rough outline, but struggle with finer object details. We replicate our experiments with two other children and find very similar results. Broadly useful high-level visual representations are thus robustly learnable from a representative sample of a child's visual experience without strong inductive biases.

READ FULL TEXT

page 5

page 7

page 9

page 10

page 18

page 19

page 22

page 25

research
02/08/2018

Learning Inductive Biases with Simple Neural Networks

People use rich prior knowledge about the world in order to efficiently ...
research
07/13/2020

Reconstruction Bottlenecks in Object-Centric Generative Models

A range of methods with suitable inductive biases exist to learn interpr...
research
07/31/2020

Self-supervised learning through the eyes of a child

Within months of birth, children have meaningful expectations about the ...
research
08/07/2023

Scaling may be all you need for achieving human-level object recognition capacity with human-like visual experience

This paper asks whether current self-supervised learning methods, if suf...
research
06/24/2019

Mutual exclusivity as a challenge for neural networks

Strong inductive biases allow children to learn in fast and adaptable wa...
research
02/16/2022

Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without Supervision

Discriminative self-supervised learning allows training models on any ra...
research
01/15/2021

Counterfactual Generative Networks

Neural networks are prone to learning shortcuts – they often model simpl...

Please sign up or login with your details

Forgot password? Click here to reset