Exploring Gradient Flow Based Saliency for DNN Model Compression

10/24/2021
by   Xinyu Liu, et al.
0

Model pruning aims to reduce the deep neural network (DNN) model size or computational overhead. Traditional model pruning methods such as l-1 pruning that evaluates the channel significance for DNN pay too much attention to the local analysis of each channel and make use of the magnitude of the entire feature while ignoring its relevance to the batch normalization (BN) and ReLU layer after each convolutional operation. To overcome these problems, we propose a new model pruning method from a new perspective of gradient flow in this paper. Specifically, we first theoretically analyze the channel's influence based on Taylor expansion by integrating the effects of BN layer and ReLU activation function. Then, the incorporation of the first-order Talyor polynomial of the scaling parameter and the shifting parameter in the BN layer is suggested to effectively indicate the significance of a channel in a DNN. Comprehensive experiments on both image classification and image denoising tasks demonstrate the superiority of the proposed novel theory and scheme. Code is available at https://github.com/CityU-AIM-Group/GFBS.

READ FULL TEXT
research
03/04/2023

Visual Saliency-Guided Channel Pruning for Deep Visual Detectors in Autonomous Driving

Deep neural network (DNN) pruning has become a de facto component for de...
research
03/15/2020

Channel Pruning Guided by Classification Loss and Feature Importance

In this work, we propose a new layer-by-layer channel pruning method cal...
research
04/06/2020

Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio

Automatic designing computationally efficient neural networks has receiv...
research
07/06/2022

Network Pruning via Feature Shift Minimization

Channel pruning is widely used to reduce the complexity of deep network ...
research
07/29/2022

Spline Representation and Redundancies of One-Dimensional ReLU Neural Network Models

We analyze the structure of a one-dimensional deep ReLU neural network (...
research
07/07/2021

Immunization of Pruning Attack in DNN Watermarking Using Constant Weight Code

To ensure protection of the intellectual property rights of DNN models, ...
research
12/05/2021

Inf-CP: A Reliable Channel Pruning based on Channel Influence

One of the most effective methods of channel pruning is to trim on the b...

Please sign up or login with your details

Forgot password? Click here to reset