Structural Compression of Convolutional Neural Networks Based on Greedy Filter Pruning

05/20/2017
by   Reza Abbasi-Asl, et al.
0

Convolutional neural networks (CNNs) have state-of-the-art performance on many problems in machine vision. However, networks with superior performance often have millions of weights so that it is difficult or impossible to use CNNs on computationally limited devices or to humanly interpret them. A myriad of CNN compression approaches have been proposed and they involve pruning and compressing the weights and filters. In this article, we introduce a greedy structural compression scheme that prunes filters in a trained CNN. We define a filter importance index equal to the classification accuracy reduction (CAR) of the network after pruning that filter (similarly defined as RAR for regression). We then iteratively prune filters based on the CAR index. This algorithm achieves substantially higher classification accuracy in AlexNet compared to other structural compression schemes that prune filters. Pruning half of the filters in the first or second layer of AlexNet, our CAR algorithm achieves 26 the best benchmark filter pruning scheme. Our CAR algorithm, combined with further weight pruning and compressing, reduces the size of first or second convolutional layer in AlexNet by a factor of 42, while achieving close to original classification accuracy through retraining (or fine-tuning) network. Finally, we demonstrate the interpretability of CAR-compressed CNNs by showing that our algorithm prunes filters with visually redundant functionalities. In fact, out of top 20 CAR-pruned filters in AlexNet, 17 of them in the first layer and 14 of them in the second layer are color-selective filters as opposed to shape-selective filters. To our knowledge, this is the first reported result on the connection between compression and interpretability of CNNs.

READ FULL TEXT

page 4

page 10

page 12

page 13

research
11/07/2017

Interpreting Convolutional Neural Networks Through Compression

Convolutional neural networks (CNNs) achieve state-of-the-art performanc...
research
03/29/2022

A Passive Similarity based CNN Filter Pruning for Efficient Acoustic Scene Classification

We present a method to develop low-complexity convolutional neural netwo...
research
07/25/2017

Towards Evolutional Compression

Compressing convolutional neural networks (CNNs) is essential for transf...
research
09/08/2020

CNNPruner: Pruning Convolutional Neural Networks with Visual Analytics

Convolutional neural networks (CNNs) have demonstrated extraordinarily g...
research
10/22/2020

Tensor Reordering for CNN Compression

We show how parameter redundancy in Convolutional Neural Network (CNN) f...
research
10/21/2019

Directed-Weighting Group Lasso for Eltwise Blocked CNN Pruning

Eltwise layer is a commonly used structure in the multi-branch deep lear...
research
01/07/2021

L2PF – Learning to Prune Faster

Various applications in the field of autonomous driving are based on con...

Please sign up or login with your details

Forgot password? Click here to reset