SimReg: Regression as a Simple Yet Effective Tool for Self-supervised Knowledge Distillation

01/13/2022
by   K L Navaneet, et al.
21

Feature regression is a simple way to distill large neural network models to smaller ones. We show that with simple changes to the network architecture, regression can outperform more complex state-of-the-art approaches for knowledge distillation from self-supervised models. Surprisingly, the addition of a multi-layer perceptron head to the CNN backbone is beneficial even if used only during distillation and discarded in the downstream task. Deeper non-linear projections can thus be used to accurately mimic the teacher without changing inference architecture and time. Moreover, we utilize independent projection heads to simultaneously distill multiple teacher networks. We also find that using the same weakly augmented image as input for both teacher and student networks aids distillation. Experiments on ImageNet dataset demonstrate the efficacy of the proposed changes in various self-supervised distillation settings.

READ FULL TEXT
research
07/21/2022

KD-MVS: Knowledge Distillation Based Self-supervised Learning for MVS

Supervised multi-view stereo (MVS) methods have achieved remarkable prog...
research
02/24/2023

Ensemble knowledge distillation of self-supervised speech models

Distilled self-supervised models have shown competitive performance and ...
research
06/13/2019

Linear Distillation Learning

Deep Linear Networks do not have expressive power but they are mathemati...
research
01/23/2023

A Simple Recipe for Competitive Low-compute Self supervised Vision Models

Self-supervised methods in vision have been mostly focused on large arch...
research
11/06/2021

Class Token and Knowledge Distillation for Multi-head Self-Attention Speaker Verification Systems

This paper explores three novel approaches to improve the performance of...
research
10/21/2022

Distilling the Undistillable: Learning from a Nasty Teacher

The inadvertent stealing of private/sensitive information using Knowledg...
research
11/01/2022

Pixel-Wise Contrastive Distillation

We present the first pixel-level self-supervised distillation framework ...

Please sign up or login with your details

Forgot password? Click here to reset