Covariate Shift in High-Dimensional Random Feature Regression

11/16/2021
by   Nilesh Tripuraneni, et al.
0

A significant obstacle in the development of robust machine learning models is covariate shift, a form of distribution shift that occurs when the input distributions of the training and test sets differ while the conditional label distributions remain the same. Despite the prevalence of covariate shift in real-world applications, a theoretical understanding in the context of modern machine learning has remained lacking. In this work, we examine the exact high-dimensional asymptotics of random feature regression under covariate shift and present a precise characterization of the limiting test error, bias, and variance in this setting. Our results motivate a natural partial order over covariate shifts that provides a sufficient condition for determining when the shift will harm (or even help) test performance. We find that overparameterized models exhibit enhanced robustness to covariate shift, providing one of the first theoretical explanations for this intriguing phenomenon. Additionally, our analysis reveals an exact linear relationship between in-distribution and out-of-distribution generalization performance, offering an explanation for this surprising recent empirical observation.

READ FULL TEXT
research
10/02/2020

Covariate Shift Adaptation in High-Dimensional and Divergent Distributions

In real world applications of supervised learning methods, training and ...
research
07/10/2023

Automatic Debiased Machine Learning for Covariate Shifts

In this paper we address the problem of bias in machine learning of para...
research
12/28/2017

Kernel Robust Bias-Aware Prediction under Covariate Shift

Under covariate shift, training (source) data and testing (target) data ...
research
11/19/2021

Maximum Mean Discrepancy for Generalization in the Presence of Distribution and Missingness Shift

Covariate shifts are a common problem in predictive modeling on real-wor...
research
12/05/2022

Blessings and Curses of Covariate Shifts: Adversarial Learning Dynamics, Directional Convergence, and Equilibria

Covariate distribution shifts and adversarial perturbations present robu...
research
05/26/2022

Undersampling is a Minimax Optimal Robustness Intervention in Nonparametric Classification

While a broad range of techniques have been proposed to tackle distribut...
research
04/18/2023

A Domain-Region Based Evaluation of ML Performance Robustness to Covariate Shift

Most machine learning methods assume that the input data distribution is...

Please sign up or login with your details

Forgot password? Click here to reset