Active Learning under Label Shift

07/16/2020
by   Eric Zhao, et al.
18

Distribution shift poses a challenge for active data collection in the real world. We address the problem of active learning under label shift and propose ALLS, the first framework for active learning under label shift. ALLS builds on label shift estimation techniques to correct for label shift with a balance of importance weighting and class-balanced sampling. We show a bias-variance trade-off between these two techniques and prove error and sample complexity bounds for a disagreement-based algorithm under ALLS. Experiments across a range of label shift settings demonstrate ALLS consistently improves performance, often reducing sample complexity by more than half an order of magnitude. Ablation studies corroborate the bias-variance trade-off revealed by our theory

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/05/2023

Active learning using region-based sampling

We present a general-purpose active learning scheme for data in metric s...
research
11/17/2022

Active Learning with Expected Error Reduction

Active learning has been studied extensively as a method for efficient d...
research
02/03/2022

GALAXY: Graph-based Active Learning at the Extreme

Active learning is a label-efficient approach to train highly effective ...
research
09/13/2021

Improving Robustness and Efficiency in Active Learning with Contrastive Loss

This paper introduces supervised contrastive active learning (SCAL) by l...
research
02/23/2016

Search Improves Label for Active Learning

We investigate active learning with access to two distinct oracles: Labe...
research
04/10/2022

Active Learning with Label Comparisons

Supervised learning typically relies on manual annotation of the true la...
research
04/05/2014

A Compression Technique for Analyzing Disagreement-Based Active Learning

We introduce a new and improved characterization of the label complexity...

Please sign up or login with your details

Forgot password? Click here to reset