Network Pruning Spaces

04/19/2023
by   Xuanyu He, et al.
0

Network pruning techniques, including weight pruning and filter pruning, reveal that most state-of-the-art neural networks can be accelerated without a significant performance drop. This work focuses on filter pruning which enables accelerated inference with any off-the-shelf deep learning library and hardware. We propose the concept of network pruning spaces that parametrize populations of subnetwork architectures. Based on this concept, we explore the structure aspect of subnetworks that result in minimal loss of accuracy in different pruning regimes and arrive at a series of observations by comparing subnetwork distributions. We conjecture through empirical studies that there exists an optimal FLOPs-to-parameter-bucket ratio related to the design of original network in a pruning regime. Statistically, the structure of a winning subnetwork guarantees an approximately optimal ratio in this regime. Upon our conjectures, we further refine the initial pruning space to reduce the cost of searching a good subnetwork architecture. Our experimental results on ImageNet show that the subnetwork we found is superior to those from the state-of-the-art pruning methods under comparable FLOPs.

READ FULL TEXT
research
09/30/2020

Pruning Filter in Filter

Pruning has become a very powerful and effective technique to compress a...
research
07/20/2017

ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression

We propose an efficient and unified framework, namely ThiNet, to simulta...
research
04/12/2022

Neural Network Pruning by Cooperative Coevolution

Neural network pruning is a popular model compression method which can s...
research
02/14/2023

Data pruning and neural scaling laws: fundamental limitations of score-based algorithms

Data pruning algorithms are commonly used to reduce the memory and compu...
research
10/14/2020

Towards Optimal Filter Pruning with Balanced Performance and Pruning Speed

Filter pruning has drawn more attention since resource constrained platf...
research
05/24/2023

SWAMP: Sparse Weight Averaging with Multiple Particles for Iterative Magnitude Pruning

Given the ever-increasing size of modern neural networks, the significan...
research
08/30/2019

Comparative study of performance of parallel Alpha Beta Pruning for different architectures

Optimization of searching the best possible action depending on various ...

Please sign up or login with your details

Forgot password? Click here to reset