Out-of-Variable Generalization

04/16/2023
by   Siyuan Guo, et al.
0

The ability of an agent to perform well in new and unseen environments is a crucial aspect of intelligence. In machine learning, this ability is referred to as strong or out-of-distribution generalization. However, simply considering differences in data distributions is not sufficient to fully capture differences in environments. In the present paper, we assay out-of-variable generalization, which refers to an agent's ability to handle new situations that involve variables never jointly observed before. We expect that such ability is important also for AI-driven scientific discovery: humans, too, explore 'Nature' by probing, observing and measuring subsets of variables at one time. Mathematically, it requires efficient re-use of past marginal knowledge, i.e., knowledge over subsets of variables. We study this problem, focusing on prediction tasks that involve observing overlapping, yet distinct, sets of causal parents. We show that the residual distribution of one environment encodes the partial derivative of the true generating function with respect to the unobserved causal parent. Hence, learning from the residual allows zero-shot prediction even when we never observe the outcome variable in the other environment.

READ FULL TEXT
research
05/11/2023

Reinterpreting causal discovery as the task of predicting unobserved joint statistics

If X,Y,Z denote sets of random variables, two different data sources may...
research
06/16/2023

BISCUIT: Causal Representation Learning from Binary Interactions

Identifying the causal variables of an environment and how to intervene ...
research
07/21/2020

Generalization and Invariances in the Presence of Unobserved Confounding

The ability to extrapolate, or generalize, from observed to new related ...
research
06/09/2022

On the Generalization and Adaption Performance of Causal Models

Learning models that offer robust out-of-distribution generalization and...
research
10/16/2012

Causal Discovery of Linear Cyclic Models from Multiple Experimental Data Sets with Overlapping Variables

Much of scientific data is collected as randomized experiments interveni...
research
02/10/2023

On the Interventional Kullback-Leibler Divergence

Modern machine learning approaches excel in static settings where a larg...
research
03/15/2022

Zipfian environments for Reinforcement Learning

As humans and animals learn in the natural world, they encounter distrib...

Please sign up or login with your details

Forgot password? Click here to reset