Learning Modular Structures That Generalize Out-of-Distribution

08/07/2022
by   Arjun Ashok, et al.
0

Out-of-distribution (O.O.D.) generalization remains to be a key challenge for real-world machine learning systems. We describe a method for O.O.D. generalization that, through training, encourages models to only preserve features in the network that are well reused across multiple training domains. Our method combines two complementary neuron-level regularizers with a probabilistic differentiable binary mask over the network, to extract a modular sub-network that achieves better O.O.D. performance than the original network. Preliminary evaluation on two benchmark datasets corroborates the promise of our method.

READ FULL TEXT

page 1

page 2

research
06/06/2022

Is a Modular Architecture Enough?

Inspired from human cognition, machine learning systems are gradually re...
research
06/09/2022

On the Generalization and Adaption Performance of Causal Models

Learning models that offer robust out-of-distribution generalization and...
research
06/05/2021

Can Subnetwork Structure be the Key to Out-of-Distribution Generalization?

Can models with particular structure avoid being biased towards spurious...
research
07/13/2020

S2RMs: Spatially Structured Recurrent Modules

Capturing the structure of a data-generating process by means of appropr...
research
04/20/2021

Gradient Matching for Domain Generalization

Machine learning systems typically assume that the distributions of trai...
research
08/20/2022

A Domain Generalization Approach for Out-Of-Distribution 12-lead ECG Classification with Convolutional Neural Networks

Deep Learning systems have achieved great success in the past few years,...
research
10/08/2021

Distinguishing rule- and exemplar-based generalization in learning systems

Despite the increasing scale of datasets in machine learning, generaliza...

Please sign up or login with your details

Forgot password? Click here to reset