Pruning Very Deep Neural Network Channels for Efficient Inference

11/14/2022
by   Yihui He, et al.
0

In this paper, we introduce a new channel pruning method to accelerate very deep convolutional neural networks. Given a trained CNN model, we propose an iterative two-step algorithm to effectively prune each layer, by a LASSO regression based channel selection and least square reconstruction. We further generalize this algorithm to multi-layer and multi-branch cases. Our method reduces the accumulated error and enhances the compatibility with various architectures. Our pruned VGG-16 achieves the state-of-the-art results by 5x speed-up along with only 0.3 is able to accelerate modern networks like ResNet, Xception and suffers only 1.4 Our code has been made publicly available.

READ FULL TEXT

page 1

page 3

page 4

research
07/19/2017

Channel Pruning for Accelerating Very Deep Neural Networks

In this paper, we introduce a new channel pruning method to accelerate v...
research
09/20/2017

Structured Probabilistic Pruning for Convolutional Neural Network Acceleration

Although deep Convolutional Neural Network (CNN) has shown better perfor...
research
05/22/2020

PruneNet: Channel Pruning via Global Importance

Channel pruning is one of the predominant approaches for accelerating de...
research
10/21/2021

CATRO: Channel Pruning via Class-Aware Trace Ratio Optimization

Deep convolutional neural networks are shown to be overkill with high pa...
research
08/06/2019

Exploiting Channel Similarity for Accelerating Deep Convolutional Neural Networks

To address the limitations of existing magnitude-based pruning algorithm...
research
10/28/2019

Layer Pruning for Accelerating Very Deep Neural Networks

In this paper, we propose an adaptive pruning method. This method can cu...
research
05/21/2020

CPOT: Channel Pruning via Optimal Transport

Recent advances in deep neural networks (DNNs) lead to tremendously grow...

Please sign up or login with your details

Forgot password? Click here to reset