Out-of-Distribution Generalization via Risk Extrapolation (REx)

03/02/2020
by   David Krueger, et al.
25

Generalizing outside of the training distribution is an open challenge for current machine learning systems. A weak form of out-of-distribution (OoD) generalization is the ability to successfully interpolate between multiple observed distributions. One way to achieve this is through robust optimization, which seeks to minimize the worst-case risk over convex combinations of the training distributions. However, a much stronger form of OoD generalization is the ability of models to extrapolate beyond the distributions observed during training. In pursuit of strong OoD generalization, we introduce the principle of Risk Extrapolation (REx). REx can be viewed as encouraging robustness over affine combinations of training risks, by encouraging strict equality between training risks. We show conceptually how this principle enables extrapolation, and demonstrate the effectiveness and scalability of instantiations of REx on various OoD generalization tasks. Our code can be found at https://github.com/capybaralet/REx_code_release.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2023

Robust Generalization against Photon-Limited Corruptions via Worst-Case Sharpness Minimization

Robust generalization aims to tackle the most challenging data distribut...
research
05/24/2023

Rethinking the Evaluation Protocol of Domain Generalization

Domain generalization aims to solve the challenge of Out-of-Distribution...
research
01/06/2022

Efficiently Disentangle Causal Representations

This paper proposes an efficient approach to learning disentangled repre...
research
02/08/2022

Uncertainty Modeling for Out-of-Distribution Generalization

Though remarkable progress has been achieved in various vision tasks, de...
research
01/16/2023

Modeling Uncertain Feature Representation for Domain Generalization

Though deep neural networks have achieved impressive success on various ...
research
01/21/2021

Out-of-Distribution Generalization Analysis via Influence Function

The mismatch between training and target data is one major challenge for...
research
11/17/2019

Encouraging an Appropriate Representation Simplifies Training of Neural Networks

A common assumption about neural networks is that they can learn an appr...

Please sign up or login with your details

Forgot password? Click here to reset