Self Training with Ensemble of Teacher Models

07/17/2021
by   Soumyadeep Ghosh, et al.
0

In order to train robust deep learning models, large amounts of labelled data is required. However, in the absence of such large repositories of labelled data, unlabeled data can be exploited for the same. Semi-Supervised learning aims to utilize such unlabeled data for training classification models. Recent progress of self-training based approaches have shown promise in this area, which leads to this study where we utilize an ensemble approach for the same. A by-product of any semi-supervised approach may be loss of calibration of the trained model especially in scenarios where unlabeled data may contain out-of-distribution samples, which leads to this investigation on how to adapt to such effects. Our proposed algorithm carefully avoids common pitfalls in utilizing unlabeled data and leads to a more accurate and calibrated supervised model compared to vanilla self-training based student-teacher algorithms. We perform several experiments on the popular STL-10 database followed by an extensive analysis of our approach and study its effects on model accuracy and calibration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2023

Enhancing Self-Training Methods

Semi-supervised learning approaches train on small sets of labeled data ...
research
04/21/2022

SelfD: Self-Learning Large-Scale Driving Policies From the Web

Effectively utilizing the vast amounts of ego-centric navigation data th...
research
08/10/2020

Knowledge Distillation and Data Selection for Semi-Supervised Learning in CTC Acoustic Models

Semi-supervised learning (SSL) is an active area of research which aims ...
research
02/23/2023

Uncertainty Guided Ensemble Self-Training for Semi-Supervised Global Field Reconstruction

Recovering a globally accurate complex physics field from limited sensor...
research
02/14/2022

Unlabeled Data Help: Minimax Analysis and Adversarial Robustness

The recent proposed self-supervised learning (SSL) approaches successful...
research
11/02/2022

More Speaking or More Speakers?

Self-training (ST) and self-supervised learning (SSL) methods have demon...
research
05/10/2019

Semi-supervised and Population Based Training for Voice Commands Recognition

We present a rapid design methodology that combines automated hyper-para...

Please sign up or login with your details

Forgot password? Click here to reset