Active Learning with Expected Error Reduction

11/17/2022
by   Stephen Mussmann, et al.
0

Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally decreases the error on an unlabeled set. However, EER requires the model to be retrained for every candidate sample and thus has not been widely used for modern deep neural networks due to this large computational cost. In this paper we reformulate EER under the lens of Bayesian active learning and derive a computationally efficient version that can use any Bayesian parameter sampling method (such as arXiv:1506.02142). We then compare the empirical performance of our method using Monte Carlo dropout for parameter sampling against state of the art methods in the deep active learning literature. Experiments are performed on four standard benchmark datasets and three WILDS datasets (arXiv:2012.07421). The results indicate that our method outperforms all other methods except one in the data shift scenario: a model dependent, non-information theoretic method that requires an order of magnitude higher computational cost (arXiv:1906.03671).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/02/2021

Efficacy of Bayesian Neural Networks in Active Learning

Obtaining labeled data for machine learning tasks can be prohibitively e...
research
06/22/2020

Effective Version Space Reduction for Convolutional Neural Networks

In active learning, sampling bias could pose a serious inconsistency pro...
research
07/16/2020

Active Learning under Label Shift

Distribution shift poses a challenge for active data collection in the r...
research
08/13/2022

Reliable emulation of complex functionals by active learning with error control

Statistical emulator is a surrogate model of complex physical models to ...
research
06/08/2023

Actively learning a Bayesian matrix fusion model with deep side information

High-dimensional deep neural network representations of images and conce...
research
06/17/2021

Gone Fishing: Neural Active Learning with Fisher Embeddings

There is an increasing need for effective active learning algorithms tha...
research
05/06/2021

Bayesian Active Learning by Disagreements: A Geometric Perspective

We present geometric Bayesian active learning by disagreements (GBALD), ...

Please sign up or login with your details

Forgot password? Click here to reset