Weighted Distillation with Unlabeled Examples

10/13/2022
by   Fotis Iliopoulos, et al.
20

Distillation with unlabeled examples is a popular and powerful method for training deep neural networks in settings where the amount of labeled data is limited: A large ”teacher” neural network is trained on the labeled data available, and then it is used to generate labels on an unlabeled dataset (typically much larger in size). These labels are then utilized to train the smaller ”student” model which will actually be deployed. Naturally, the success of the approach depends on the quality of the teacher's labels, since the student could be confused if trained on inaccurate data. This paper proposes a principled approach for addressing this issue based on a ”debiasing” reweighting of the student's loss function tailored to the distillation training paradigm. Our method is hyper-parameter free, data-agnostic, and simple to implement. We demonstrate significant improvements on popular academic datasets and we accompany our results with a theoretical analysis which rigorously justifies the performance of our method in certain settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2023

SLaM: Student-Label Mixing for Semi-Supervised Knowledge Distillation

Semi-supervised knowledge distillation is a powerful training paradigm f...
research
05/22/2021

Revisiting Knowledge Distillation for Object Detection

The existing solutions for object detection distillation rely on the ava...
research
02/13/2021

Distilling Double Descent

Distillation is the technique of training a "student" model based on exa...
research
10/04/2019

Distilling Transformers into Simple Neural Networks with Unlabeled Transfer Data

Recent advances in pre-training huge models on large amounts of text thr...
research
06/21/2018

Learning to Rank from Samples of Variable Quality

Training deep neural networks requires many training samples, but in pra...
research
07/25/2019

Towards Generalizing Sensorimotor Control Across Weather Conditions

The ability of deep learning models to generalize well across different ...
research
11/08/2017

Fidelity-Weighted Learning

Training deep neural networks requires many training samples, but in pra...

Please sign up or login with your details

Forgot password? Click here to reset