Semi-Supervised Active Learning with Temporal Output Discrepancy

07/29/2021
by   Siyu Huang, et al.
0

While deep learning succeeds in a wide range of tasks, it highly depends on the massive collection of annotated data which is expensive and time-consuming. To lower the cost of data annotation, active learning has been proposed to interactively query an oracle to annotate a small proportion of informative samples in an unlabeled dataset. Inspired by the fact that the samples with higher loss are usually more informative to the model than the samples with lower loss, in this paper we present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss. The core of our approach is a measurement Temporal Output Discrepancy (TOD) that estimates the sample loss by evaluating the discrepancy of outputs given by models at different optimization steps. Our theoretical investigation shows that TOD lower-bounds the accumulated sample loss thus it can be used to select informative unlabeled samples. On basis of TOD, we further develop an effective unlabeled data sampling strategy as well as an unsupervised learning criterion that enhances model performance by incorporating the unlabeled data. Due to the simplicity of TOD, our active learning approach is efficient, flexible, and task-agnostic. Extensive experimental results demonstrate that our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/20/2022

Temporal Output Discrepancy for Loss Estimation-based Active Learning

While deep learning succeeds in a wide range of tasks, it highly depends...
research
08/20/2021

Influence Selection for Active Learning

The existing active learning methods select the samples by evaluating th...
research
06/23/2017

A Variance Maximization Criterion for Active Learning

Active learning aims to train a classifier as fast as possible with as f...
research
01/19/2022

Using Self-Supervised Pretext Tasks for Active Learning

Labeling a large set of data is expensive. Active learning aims to tackl...
research
12/10/2021

Boosting Active Learning via Improving Test Performance

Central to active learning (AL) is what data should be selected for anno...
research
01/22/2021

DSAL: Deeply Supervised Active Learning from Strong and Weak Labelers for Biomedical Image Segmentation

Image segmentation is one of the most essential biomedical image process...
research
07/16/2021

The Application of Active Query K-Means in Text Classification

Active learning is a state-of-art machine learning approach to deal with...

Please sign up or login with your details

Forgot password? Click here to reset