On Discarding, Caching, and Recalling Samples in Active Learning

06/20/2012
by   Ashish Kapoor, et al.
0

We address challenges of active learning under scarce informational resources in non-stationary environments. In real-world settings, data labeled and integrated into a predictive model may become invalid over time. However, the data can become informative again with switches in context and such changes may indicate unmodeled cyclic or other temporal dynamics. We explore principles for discarding, caching, and recalling labeled data points in active learning based on computations of value of information. We review key concepts and study the value of the methods via investigations of predictive performance and costs of acquiring data for simulated and real-world data sets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/23/2021

One-Round Active Learning

Active learning has been a main solution for reducing data labeling cost...
research
04/06/2021

Low-Regret Active learning

We develop an online learning algorithm for identifying unlabeled data p...
research
10/27/2020

Graph-based Reinforcement Learning for Active Learning in Real Time: An Application in Modeling River Networks

Effective training of advanced ML models requires large amounts of label...
research
02/08/2023

Combining self-labeling and demand based active learning for non-stationary data streams

Learning from non-stationary data streams is a research direction that g...
research
02/18/2020

Active Learning-based Classification in Automated Connected Vehicles

Machine learning has emerged as a promising paradigm for enabling connec...
research
10/04/2022

Active Learning for Regression with Aggregated Outputs

Due to the privacy protection or the difficulty of data collection, we c...
research
04/17/2023

Prediction-Oriented Bayesian Active Learning

Information-theoretic approaches to active learning have traditionally f...

Please sign up or login with your details

Forgot password? Click here to reset