Understanding Diversity based Pruning of Neural Networks – Statistical Mechanical Analysis

06/30/2020
by   Rupam Acharyya, et al.
12

Deep learning architectures with a huge number of parameters are often compressed using pruning techniques to ensure computational efficiency of inference during deployment. Despite multitude of empirical advances, there is no theoretical understanding of the effectiveness of different pruning methods. We address this issue by setting up the problem in the statistical mechanics formulation of a teacher-student framework and deriving generalization error (GE) bounds of specific pruning methods. This theoretical premise allows comparison between pruning methods and we use it to investigate compression of neural networks via diversity-based pruning methods. A recent work showed that Determinantal Point Process (DPP) based node pruning method is notably superior to competing approaches when tested on real datasets. Using GE bounds in the aforementioned setup we provide theoretical guarantees for their empirical observations. Another consistent finding in literature is that sparse neural networks (edge pruned) generalize better than dense neural networks (node pruned) for a fixed number of parameters. We use our theoretical setup to prove that baseline random edge pruning method performs better than DPP node pruning method. Finally, we draw motivation from our theoretical results to propose a DPP edge pruning technique for neural networks which empirically outperforms other competing pruning methods on real datasets.

READ FULL TEXT

page 15

page 20

page 21

research
05/23/2021

Spectral Pruning for Recurrent Neural Networks

Pruning techniques for neural networks with a recurrent architecture, su...
research
09/21/2020

Conditional Automated Channel Pruning for Deep Neural Networks

Model compression aims to reduce the redundancy of deep networks to obta...
research
03/12/2018

FeTa: A DCA Pruning Algorithm with Generalization Error Guarantees

Recent DNN pruning algorithms have succeeded in reducing the number of p...
research
10/24/2019

A Comparative Study of Neural Network Compression

There has recently been an increasing desire to evaluate neural networks...
research
06/21/2023

Quantifying lottery tickets under label noise: accuracy, calibration, and complexity

Pruning deep neural networks is a widely used strategy to alleviate the ...
research
12/03/2018

Structure Learning Using Forced Pruning

Markov networks are widely used in many Machine Learning applications in...
research
05/30/2023

Generalization Bounds for Magnitude-Based Pruning via Sparse Matrix Sketching

In this paper, we derive a novel bound on the generalization error of Ma...

Please sign up or login with your details

Forgot password? Click here to reset