Moonshine: Distilling with Cheap Convolutions

11/07/2017
by   Elliot J. Crowley, et al.
0

Model distillation compresses a trained machine learning model, such as a neural network, into a smaller alternative such that it could be easily deployed in a resource limited setting. Unfortunately, this requires engineering two architectures: a student architecture smaller than the first teacher architecture but trained to emulate it. In this paper, we present a distillation strategy that produces a student architecture that is a simple transformation of the teacher architecture. Recent model distillation methods allow us to preserve most of the performance of the trained model after replacing convolutional blocks with a cheap alternative. In addition, distillation by attention transfer provides student network performance that is better than training that student architecture directly on data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/03/2020

Interactive Knowledge Distillation

Knowledge distillation is a standard teacher-student learning framework ...
research
02/10/2020

Subclass Distillation

After a large "teacher" neural network has been trained on labeled data,...
research
01/29/2023

Pipe-BD: Pipelined Parallel Blockwise Distillation

Training large deep neural network models is highly challenging due to t...
research
04/04/2023

Label-guided Attention Distillation for Lane Segmentation

Contemporary segmentation methods are usually based on deep fully convol...
research
08/23/2020

Matching Guided Distillation

Feature distillation is an effective way to improve the performance for ...
research
01/26/2018

Transparent Model Distillation

Model distillation was originally designed to distill knowledge from a l...
research
05/31/2019

The Pupil Has Become the Master: Teacher-Student Model-Based Word Embedding Distillation with Ensemble Learning

Recent advances in deep learning have facilitated the demand of neural m...

Please sign up or login with your details

Forgot password? Click here to reset