BREEDS: Benchmarks for Subpopulation Shift

08/11/2020
by   Shibani Santurkar, et al.
5

We develop a methodology for assessing the robustness of models to subpopulation shift—specifically, their ability to generalize to novel data subpopulations that were not observed during training. Our approach leverages the class structure underlying existing datasets to control the data subpopulations that comprise the training and test distributions. This enables us to synthesize realistic distribution shifts whose sources can be precisely controlled and characterized, within existing large-scale datasets. Applying this methodology to the ImageNet dataset, we create a suite of subpopulation shift benchmarks of varying granularity. We then validate that the corresponding shifts are tractable by obtaining human baselines for them. Finally, we utilize these benchmarks to measure the sensitivity of standard model architectures as well as the effectiveness of off-the-shelf train-time robustness interventions. Code and data available at https://github.com/MadryLab/BREEDS-Benchmarks .

READ FULL TEXT

page 6

page 23

page 25

page 27

page 28

page 29

page 31

research
08/30/2021

SHIFT15M: Multiobjective Large-Scale Fashion Dataset with Distributional Shifts

Many machine learning algorithms assume that the training data and the t...
research
02/15/2023

Dataset Interfaces: Diagnosing Model Failures Using Controllable Counterfactual Generation

Distribution shifts are a major source of failure of deployed machine le...
research
02/23/2023

Change is Hard: A Closer Look at Subpopulation Shift

Machine learning models often perform poorly on subgroups that are under...
research
03/27/2023

GeoNet: Benchmarking Unsupervised Adaptation across Geographies

In recent years, several efforts have been aimed at improving the robust...
research
08/07/2023

Distributionally Robust Classification on a Data Budget

Real world uses of deep learning require predictable model behavior unde...
research
06/30/2022

AnoShift: A Distribution Shift Benchmark for Unsupervised Anomaly Detection

Analyzing the distribution shift of data is a growing research direction...
research
02/05/2023

Leaving Reality to Imagination: Robust Classification via Generated Datasets

Recent research on robustness has revealed significant performance gaps ...

Please sign up or login with your details

Forgot password? Click here to reset