Conditional Automated Channel Pruning for Deep Neural Networks

09/21/2020
by   Yixin Liu, et al.
35

Model compression aims to reduce the redundancy of deep networks to obtain compact models. Recently, channel pruning has become one of the predominant compression methods to deploy deep models on resource-constrained devices. Most channel pruning methods often use a fixed compression rate for all the layers of the model, which, however, may not be optimal. To address this issue, given a target compression rate for the whole model, one can search for the optimal compression rate for each layer. Nevertheless, these methods perform channel pruning for a specific target compression rate. When we consider multiple compression rates, they have to repeat the channel pruning process multiple times, which is very inefficient yet unnecessary. To address this issue, we propose a Conditional Automated Channel Pruning(CACP) method to obtain the compressed models with different compression rates through single channel pruning process. To this end, we develop a conditional model that takes an arbitrary compression rate as input and outputs the corresponding compressed model. In the experiments, the resultant models with different compression rates consistently outperform the models compressed by existing methods with a channel pruning process for each target compression rate.

READ FULL TEXT

page 1

page 2

research
06/14/2018

PCAS: Pruning Channels with Attention Statistics

To implement deep neural networks on small embedded devices, conventiona...
research
05/29/2018

A novel channel pruning method for deep neural network compression

In recent years, deep neural networks have achieved great success in the...
research
04/25/2023

Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures

Pruning - that is, setting a significant subset of the parameters of a n...
research
02/19/2020

Model-Agnostic Structured Sparsification with Learnable Channel Shuffle

Recent advances in convolutional neural networks (CNNs) usually come wit...
research
04/30/2021

Stealthy Backdoors as Compression Artifacts

In a backdoor attack on a machine learning model, an adversary produces ...
research
09/07/2022

Interpretations Steered Network Pruning via Amortized Inferred Saliency Maps

Convolutional Neural Networks (CNNs) compression is crucial to deploying...
research
06/30/2020

Understanding Diversity based Pruning of Neural Networks – Statistical Mechanical Analysis

Deep learning architectures with a huge number of parameters are often c...

Please sign up or login with your details

Forgot password? Click here to reset