Online Ensemble Model Compression using Knowledge Distillation

11/15/2020
by   Devesh Walawalkar, et al.
0

This paper presents a novel knowledge distillation based model compression framework consisting of a student ensemble. It enables distillation of simultaneously learnt ensemble knowledge onto each of the compressed student models. Each model learns unique representations from the data distribution due to its distinct architecture. This helps the ensemble generalize better by combining every model's knowledge. The distilled students and ensemble teacher are trained simultaneously without requiring any pretrained weights. Moreover, our proposed method can deliver multi-compressed students with single training, which is efficient and flexible for different scenarios. We provide comprehensive experiments using state-of-the-art classification models to validate our framework's effectiveness. Notably, using our framework a 97 compressed ResNet110 student model managed to produce a 10.64 accuracy gain over its individual baseline training on CIFAR100 dataset. Similarly a 95 accuracy gain.

READ FULL TEXT
research
04/18/2023

Deep Collective Knowledge Distillation

Many existing studies on knowledge distillation have focused on methods ...
research
10/21/2021

Augmenting Knowledge Distillation With Peer-To-Peer Mutual Learning For Model Compression

Knowledge distillation (KD) is an effective model compression technique ...
research
09/24/2019

FEED: Feature-level Ensemble for Knowledge Distillation

Knowledge Distillation (KD) aims to transfer knowledge in a teacher-stud...
research
07/03/2020

Knowledge Distillation Beyond Model Compression

Knowledge distillation (KD) is commonly deemed as an effective model com...
research
01/31/2018

Model compression for faster structural separation of macromolecules captured by Cellular Electron Cryo-Tomography

Electron Cryo-Tomography (ECT) enables 3D visualization of macromolecule...
research
12/06/2022

Leveraging Different Learning Styles for Improved Knowledge Distillation

Learning style refers to a type of training mechanism adopted by an indi...
research
03/28/2023

Enhancing Depth Completion with Multi-View Monitored Distillation

This paper presents a novel method for depth completion, which leverages...

Please sign up or login with your details

Forgot password? Click here to reset