Log In Sign Up

Learning in Confusion: Batch Active Learning with Noisy Oracle

We study the problem of training machine learning models incrementally using active learning with access to imperfect or noisy oracles. We specifically consider the setting of batch active learning, in which multiple samples are selected as opposed to a single sample as in classical settings so as to reduce the training overhead. Our approach bridges between uniform randomness and score based importance sampling of clusters when selecting a batch of new samples. Experiments on benchmark image classification datasets (MNIST, SVHN, and CIFAR10) shows improvement over existing active learning strategies. We introduce an extra denoising layer to deep networks to make active learning robust to label noises and show significant improvements.


page 1

page 2

page 3

page 4


Diverse mini-batch Active Learning

We study the problem of reducing the amount of labeled training data req...

Batch Active Learning at Scale

The ability to train complex and highly effective models often requires ...

The Practical Challenges of Active Learning: Lessons Learned from Live Experimentation

We tested in a live setting the use of active learning for selecting tex...

Active Learning for Single Neuron Models with Lipschitz Non-Linearities

We consider the problem of active learning for single neuron models, als...

On the reusability of samples in active learning

An interesting but not extensively studied question in active learning i...

Search Improves Label for Active Learning

We investigate active learning with access to two distinct oracles: Labe...

Minimum-Margin Active Learning

We present a new active sampling method we call min-margin which trains ...