Distilling with Performance Enhanced Students

10/24/2018
by   Jack Turner, et al.
0

The task of accelerating large neural networks on general purpose hardware has, in recent years, prompted the use of channel pruning to reduce network size. However, the efficacy of pruning based approaches has since been called into question. In this paper, we turn to distillation for model compression---specifically, attention transfer---and develop a simple method for discovering performance enhanced student networks. We combine channel saliency metrics with empirical observations of runtime performance to design more accurate networks for a given latency budget. We apply our methodology to residual and densely-connected networks, and show that we are able to find resource-efficient student networks on different hardware platforms while maintaining very high accuracy. These performance-enhanced student networks achieve up to 10 counterparts for the same inference time.

READ FULL TEXT
research
10/24/2018

HAKD: Hardware Aware Knowledge Distillation

Despite recent developments, deploying deep neural networks on resource ...
research
10/14/2022

Lightweight Alpha Matting Network Using Distillation-Based Channel Pruning

Recently, alpha matting has received a lot of attention because of its u...
research
03/04/2023

Visual Saliency-Guided Channel Pruning for Deep Visual Detectors in Autonomous Driving

Deep neural network (DNN) pruning has become a de facto component for de...
research
10/13/2022

Structural Pruning via Latency-Saliency Knapsack

Structural pruning can simplify network architecture and improve inferen...
research
02/14/2021

ChipNet: Budget-Aware Pruning with Heaviside Continuous Approximations

Structured pruning methods are among the effective strategies for extrac...
research
09/30/2021

Deep Neural Compression Via Concurrent Pruning and Self-Distillation

Pruning aims to reduce the number of parameters while maintaining perfor...
research
06/29/2022

Weighted ensemble: Recent mathematical developments

The weighted ensemble (WE) method, an enhanced sampling approach based o...

Please sign up or login with your details

Forgot password? Click here to reset