Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers

02/01/2018
by   Jianbo Ye, et al.
0

Model pruning has become a useful technique that improves the computational efficiency of deep learning, making it possible to deploy solutions on resource-limited scenarios. A widely-used practice in relevant work assumes that a smaller-norm parameter or feature plays a less informative role at the inference time. In this paper, we propose a channel pruning technique for accelerating the computations of deep convolutional neural networks (CNNs), which does not critically rely on this assumption. Instead, it focuses on direct simplification of the channel-to-channel computation graph of a CNN without the need of performing a computational difficult and not always useful task of making high-dimensional tensors of CNN structured sparse. Our approach takes two stages: the first being to adopt an end-to-end stochastic training method that eventually forces the outputs of some channels being constant, and the second being to prune those constant channels from the original neural network by adjusting the biases of their impacting layers such that the resulting compact model can be quickly fine-tuned. Our approach is mathematically appealing from an optimization perspective and easy to reproduce. We experimented our approach through several image learning benchmarks and demonstrate its interesting aspects and the competitive performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/02/2021

Network Pruning via Resource Reallocation

Channel pruning is broadly recognized as an effective approach to obtain...
research
10/03/2020

UCP: Uniform Channel Pruning for Deep Convolutional Neural Networks Compression and Acceleration

To apply deep CNNs to mobile terminals and portable devices, many schola...
research
11/12/2020

When deep learning models on GPU can be accelerated by taking advantage of unstructured sparsity

This paper is focused on the improvement the efficiency of the sparse co...
research
10/12/2018

Dynamic Channel Pruning: Feature Boosting and Suppression

Making deep convolutional neural networks more accurate typically comes ...
research
05/08/2021

EZCrop: Energy-Zoned Channels for Robust Output Pruning

Recent results have revealed an interesting observation in a trained con...
research
05/15/2019

Dynamic Neural Network Channel Execution for Efficient Training

Existing methods for reducing the computational burden of neural network...
research
11/23/2020

End-to-End Framework for Efficient Deep Learning Using Metasurfaces Optics

Deep learning using Convolutional Neural Networks (CNNs) has been shown ...

Please sign up or login with your details

Forgot password? Click here to reset