One-Round Active Learning

04/23/2021
by   Tianhao Wang, et al.
0

Active learning has been a main solution for reducing data labeling costs. However, existing active learning strategies assume that a data owner can interact with annotators in an online, timely manner, which is usually impractical. Even with such interactive annotators, for existing active learning strategies to be effective, they often require many rounds of interactions between the data owner and annotators, which is often time-consuming. In this work, we initiate the study of one-round active learning, which aims to select a subset of unlabeled data points that achieve the highest utility after being labeled with only the information from initially labeled data points. We propose DULO, a general framework for one-round active learning based on the notion of data utility functions, which map a set of data points to some performance measure of the model trained on the set. We formulate the one-round active learning problem as data utility function maximization. We further propose strategies to make the estimation and optimization of data utility functions scalable to large models and large unlabeled data sets. Our results demonstrate that while existing active learning approaches could succeed with multiple rounds, DULO consistently performs better in the one-round setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2021

Zero-Round Active Learning

Active learning (AL) aims at reducing labeling effort by identifying the...
research
10/10/2019

Active Learning with Importance Sampling

We consider an active learning setting where the algorithm has access to...
research
02/17/2023

A survey on online active learning

Online active learning is a paradigm in machine learning that aims to se...
research
04/06/2021

Low-Regret Active learning

We develop an online learning algorithm for identifying unlabeled data p...
research
05/23/2020

Active Learning for Skewed Data Sets

Consider a sequential active learning problem where, at each round, an a...
research
06/20/2012

On Discarding, Caching, and Recalling Samples in Active Learning

We address challenges of active learning under scarce informational reso...
research
06/06/2023

Query Complexity of Active Learning for Function Family With Nearly Orthogonal Basis

Many machine learning algorithms require large numbers of labeled data t...

Please sign up or login with your details

Forgot password? Click here to reset