Bitrate-Constrained DRO: Beyond Worst Case Robustness To Unknown Group Shifts

02/06/2023
by   Amrith Setlur, et al.
0

Training machine learning models robust to distribution shifts is critical for real-world applications. Some robust training algorithms (e.g., Group DRO) specialize to group shifts and require group information on all training points. Other methods (e.g., CVaR DRO) that do not need group annotations can be overly conservative, since they naively upweight high loss points which may form a contrived set that does not correspond to any meaningful group in the real world (e.g., when the high loss points are randomly mislabeled training points). In this work, we address limitations in prior approaches by assuming a more nuanced form of group shift: conditioned on the label, we assume that the true group function (indicator over group) is simple. For example, we may expect that group shifts occur along low bitrate features (e.g., image background, lighting). Thus, we aim to learn a model that maintains high accuracy on simple group functions realized by these low bitrate features, that need not spend valuable model capacity achieving high accuracy on contrived groups of examples. Based on this, we consider the two-player game formulation of DRO where the adversary's capacity is bitrate-constrained. Our resulting practical algorithm, Bitrate-Constrained DRO (BR-DRO), does not require group information on training samples yet matches the performance of Group DRO on datasets that have training group annotations and that of CVaR DRO on long-tailed distributions. Our theoretical analysis reveals that in some settings BR-DRO objective can provably yield statistically efficient and less conservative solutions than unconstrained CVaR DRO.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/19/2021

Just Train Twice: Improving Group Robustness without Training Group Information

Standard training via empirical risk minimization (ERM) can produce mode...
research
03/29/2021

Learning Under Adversarial and Interventional Shifts

Machine learning models are often trained on data from one distribution ...
research
10/18/2018

Removing the influence of a group variable in high-dimensional predictive modelling

Predictive modelling relies on the assumption that observations used for...
research
02/23/2023

Change is Hard: A Closer Look at Subpopulation Shift

Machine learning models often perform poorly on subgroups that are under...
research
08/26/2022

Take One Gram of Neural Features, Get Enhanced Group Robustness

Predictive performance of machine learning models trained with empirical...
research
05/25/2023

Rectifying Group Irregularities in Explanations for Distribution Shift

It is well-known that real-world changes constituting distribution shift...
research
05/26/2022

Undersampling is a Minimax Optimal Robustness Intervention in Nonparametric Classification

While a broad range of techniques have been proposed to tackle distribut...

Please sign up or login with your details

Forgot password? Click here to reset