MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks

11/21/2019
by   Yunteng Luan, et al.
0

As the development of neural networks, more and more deep neural networks are adopted in various tasks, such as image classification. However, as the huge computational overhead, these networks could not be applied on mobile devices or other low latency scenes. To address this dilemma, muti-exit convolutional network is proposed to allow faster inference via early exits with the corresponding classifiers. These networks utilize sophisticated designing to increase the early exit accuracy. However, naively training the multi-exit network could hurt the performance (accuracy) of deep neural networks as early-exit classifiers throughout interfere with the feature generation process. In this paper, we propose a general training framework named multi-self-distillation learning (MSD), which mining knowledge of different classifiers within the same network and boost every classifier accuracy. Our approach can be applied not only to multi-exit networks, but also modern CNNs (e.g., ResNet Series) augmented with additional side branch classifiers. We use sampling-based branch augmentation technique to transform a single-exit network into a multi-exit network. This reduces the gap of capacity between different classifiers, and improves the effectiveness of applying MSD. Our experiments show that MSD improves the accuracy of various networks: enhancing the accuracy of every classifier significantly for existing multi-exit network (MSDNet), improving vanilla single-exit networks with internal classifiers with high accuracy, while also improving the final accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/17/2019

Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation

Convolutional neural networks have been widely deployed in various appli...
research
09/06/2017

BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks

Deep neural networks are state of the art methods for many learning task...
research
01/14/2021

A Multiple Classifier Approach for Concatenate-Designed Neural Networks

This article introduces a multiple classifier method to improve the perf...
research
01/28/2023

Anticipate, Ensemble and Prune: Improving Convolutional Neural Networks via Aggregated Early Exits

Today, artificial neural networks are the state of the art for solving a...
research
07/14/2022

T-RECX: Tiny-Resource Efficient Convolutional Neural Networks with Early-Exit

Deploying Machine learning (ML) on the milliwatt-scale edge devices (tin...
research
04/04/2022

Efficient, Uncertainty-based Moderation of Neural Networks Text Classifiers

To maximize the accuracy and increase the overall acceptance of text cla...
research
04/21/2021

Improving the Accuracy of Early Exits in Multi-Exit Architectures via Curriculum Learning

Deploying deep learning services for time-sensitive and resource-constra...

Please sign up or login with your details

Forgot password? Click here to reset