Pruning by Explaining: A Novel Criterion for Deep Neural Network Pruning

12/18/2019
by   Seul-Ki Yeom, et al.
68

The success of convolutional neural networks (CNNs) in various applications is accompanied by a significant increase in computation and parameter storage costs. Recent efforts to reduce these overheads involve pruning and compressing the weights of various layers while at the same time aiming to not sacrifice performance. In this paper, we propose a novel criterion for CNN pruning inspired by neural network interpretability: The most relevant elements, i.e. weights or filters, are automatically found using their relevance score in the sense of explainable AI (XAI). By that we for the first time link the two disconnected lines of interpretability and model compression research. We show in particular that our proposed method can efficiently prune transfer-learned CNN models where networks pre-trained on large corpora are adapted to specialized tasks. To this end, the method is evaluated on a broad range of computer vision datasets. Notably, our novel criterion is not only competitive or better compared to state-of-the-art pruning criteria when successive retraining is performed, but clearly outperforms these previous criteria in the common application setting where the data of the task to be transferred to are very scarce and no retraining is possible. Our method can iteratively compress the model while maintaining or even improving accuracy. At the same time, it has a computational cost in the order of gradient computation and is comparatively simple to apply without the need for tuning hyperparameters for pruning.

READ FULL TEXT

page 7

page 10

page 16

research
09/20/2017

Structured Probabilistic Pruning for Convolutional Neural Network Acceleration

Although deep Convolutional Neural Network (CNN) has shown better perfor...
research
08/31/2016

Pruning Filters for Efficient ConvNets

The success of CNNs in various applications is accompanied by a signific...
research
01/12/2021

Convolutional Neural Network Simplification with Progressive Retraining

Kernel pruning methods have been proposed to speed up, simplify, and imp...
research
02/17/2023

Less is More: The Influence of Pruning on the Explainability of CNNs

Modern, state-of-the-art Convolutional Neural Networks (CNNs) in compute...
research
11/19/2016

Pruning Convolutional Neural Networks for Resource Efficient Inference

We propose a new formulation for pruning convolutional kernels in neural...
research
03/08/2023

X-Pruner: eXplainable Pruning for Vision Transformers

Recently vision transformer models have become prominent models for a ra...
research
04/25/2023

Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures

Pruning - that is, setting a significant subset of the parameters of a n...

Please sign up or login with your details

Forgot password? Click here to reset