Out-distribution aware Self-training in an Open World Setting

12/21/2020
by   Maximilian Augustin, et al.
23

Deep Learning heavily depends on large labeled datasets which limits further improvements. While unlabeled data is available in large amounts, in particular in image recognition, it does not fulfill the closed world assumption of semi-supervised learning that all unlabeled data are task-related. The goal of this paper is to leverage unlabeled data in an open world setting to further improve prediction performance. For this purpose, we introduce out-distribution aware self-training, which includes a careful sample selection strategy based on the confidence of the classifier. While normal self-training deteriorates prediction performance, our iterative scheme improves using up to 15 times the amount of originally labeled data. Moreover, our classifiers are by design out-distribution aware and can thus distinguish task-related inputs from unrelated ones.

READ FULL TEXT

page 14

page 18

page 19

page 20

page 21

page 23

page 26

page 28

research
04/29/2022

OSSGAN: Open-Set Semi-Supervised Image Generation

We introduce a challenging training scheme of conditional GANs, called o...
research
12/13/2020

Improving the Classification of Rare Chords with Unlabeled Data

In this work, we explore techniques to improve performance for rare clas...
research
04/21/2022

SelfD: Self-Learning Large-Scale Driving Policies From the Web

Effectively utilizing the vast amounts of ego-centric navigation data th...
research
10/22/2019

Class Mean Vectors, Self Monitoring and Self Learning for Neural Classifiers

In this paper we explore the role of sample mean in building a neural ne...
research
12/09/2021

Extending the WILDS Benchmark for Unsupervised Adaptation

Machine learning systems deployed in the wild are often trained on a sou...
research
01/13/2019

Gradient Regularized Budgeted Boosting

As machine learning transitions increasingly towards real world applicat...
research
09/30/2019

Revisiting Self-Training for Neural Sequence Generation

Self-training is one of the earliest and simplest semi-supervised method...

Please sign up or login with your details

Forgot password? Click here to reset