Distinguishing rule- and exemplar-based generalization in learning systems

10/08/2021
by   Ishita Dasgupta, et al.
0

Despite the increasing scale of datasets in machine learning, generalization to unseen regions of the data distribution remains crucial. Such extrapolation is by definition underdetermined and is dictated by a learner's inductive biases. Machine learning systems often do not share the same inductive biases as humans and, as a result, extrapolate in ways that are inconsistent with our expectations. We investigate two distinct such inductive biases: feature-level bias (differences in which features are more readily learned) and exemplar-vs-rule bias (differences in how these learned features are used for generalization). Exemplar- vs. rule-based generalization has been studied extensively in cognitive psychology, and, in this work, we present a protocol inspired by these experimental approaches for directly probing this trade-off in learning systems. The measures we propose characterize changes in extrapolation behavior when feature coverage is manipulated in a combinatorial setting. We present empirical results across a range of models and across both expository and real-world image and language domains. We demonstrate that measuring the exemplar-rule trade-off while controlling for feature-level bias provides a more complete picture of extrapolation behavior than existing formalisms. We find that most standard neural network models have a propensity towards exemplar-based extrapolation and discuss the implications of these findings for research on data augmentation, fairness, and systematic generalization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/13/2023

Data Augmentation via Subgroup Mixup for Improving Fairness

In this work, we propose data augmentation via pairwise mixup across sub...
research
04/09/2018

A review of possible effects of cognitive biases on interpretation of rule-based machine learning models

This paper investigates to what extent do cognitive biases affect human ...
research
02/10/2022

Deconstructing the Inductive Biases of Hamiltonian Neural Networks

Physics-inspired neural networks (NNs), such as Hamiltonian or Lagrangia...
research
10/11/2022

Transformers generalize differently from information stored in context vs in weights

Transformer models can use two fundamentally different kinds of informat...
research
06/05/2023

Learning to Substitute Spans towards Improving Compositional Generalization

Despite the rising prevalence of neural sequence models, recent empirica...
research
08/07/2022

Learning Modular Structures That Generalize Out-of-Distribution

Out-of-distribution (O.O.D.) generalization remains to be a key challeng...
research
10/12/2021

Trivial or impossible – dichotomous data difficulty masks model differences (on ImageNet and beyond)

"The power of a generalization system follows directly from its biases" ...

Please sign up or login with your details

Forgot password? Click here to reset