PAC Generalization via Invariant Representations

05/30/2022
by   Advait Parulekar, et al.
0

One method for obtaining generalizable solutions to machine learning tasks when presented with diverse training environments is to find invariant representations of the data. These are representations of the covariates such that the best model on top of the representation is invariant across training environments. In the context of linear Structural Equation Models (SEMs), invariant representations might allow us to learn models with out-of-distribution guarantees, i.e., models that are robust to interventions in the SEM. To address the invariant representation problem in a finite sample setting, we consider the notion of ϵ-approximate invariance. We study the following question: If a representation is approximately invariant with respect to a given number of training interventions, will it continue to be approximately invariant on a larger collection of unseen SEMs? This larger collection of SEMs is generated through a parameterized family of interventions. Inspired by PAC learning, we obtain finite-sample out-of-distribution generalization guarantees for approximate invariance that holds probabilistically over a family of linear SEMs without faithfulness assumptions. Our results show bounds that do not scale in ambient dimension when intervention sites are restricted to lie in a constant size subset of in-degree bounded nodes. We also show how to extend our results to a linear indirect observation model that incorporates latent variables.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/22/2022

Learning Invariant Representations under General Interventions on the Response

It has become increasingly common nowadays to collect observations of fe...
research
04/07/2023

Domain Generalization In Robust Invariant Representation

Unsupervised approaches for learning representations invariant to common...
research
06/07/2022

Distributionally Invariant Learning: Rationalization and Practical Algorithms

The invariance property across environments is at the heart of invariant...
research
06/12/2020

The Difficult Task of Distribution Generalization in Nonlinear Models

We consider the problem of predicting a response from a set of covariate...
research
11/10/2021

Understanding the Generalization Benefit of Model Invariance from a Data Perspective

Machine learning models that are developed to be invariant under certain...
research
06/23/2022

Invariant Causal Mechanisms through Distribution Matching

Learning representations that capture the underlying data generating pro...
research
01/30/2023

Bagging Provides Assumption-free Stability

Bagging is an important technique for stabilizing machine learning model...

Please sign up or login with your details

Forgot password? Click here to reset