Structured Pruning of Neural Networks with Budget-Aware Regularization

11/23/2018
by   Carl Lemaire, et al.
0

Pruning methods have shown to be effective at reducing the size of deep neural networks while keeping accuracy almost intact. Among the most effective methods are those that prune a network while training it with a sparsity prior loss and learnable dropout parameters. A shortcoming of these approaches however is that neither the size nor the inference speed of the pruned network can be controlled directly; yet this is a key feature for targeting deployment of CNNs on low-power hardware. To overcome this, we introduce a budgeted regularized pruning framework for deep convolutional neural networks. Our approach naturally fits into traditional neural network training as it consists of a learnable masking layer, a novel budget-aware objective function, and the use of knowledge distillation. We also provide insights on how to prune a residual network and how this can lead to new architectures. Experimental results reveal that CNNs pruned with our method are more accurate and less compute-hungry than state-of-the-art methods. Also, our approach is more effective at preventing accuracy collapse in case of severe pruning; this allows us to attain pruning factors up to 16x without significantly affecting the accuracy.

READ FULL TEXT
research
07/14/2022

PR-DARTS: Pruning-Based Differentiable Architecture Search

The deployment of Convolutional Neural Networks (CNNs) on edge devices i...
research
10/06/2020

Comprehensive Online Network Pruning via Learnable Scaling Factors

One of the major challenges in deploying deep neural network architectur...
research
02/14/2021

ChipNet: Budget-Aware Pruning with Heaviside Continuous Approximations

Structured pruning methods are among the effective strategies for extrac...
research
08/04/2023

Pruning a neural network using Bayesian inference

Neural network pruning is a highly effective technique aimed at reducing...
research
09/10/2021

On the Compression of Neural Networks Using ℓ_0-Norm Regularization and Weight Pruning

Despite the growing availability of high-capacity computational platform...
research
05/18/2017

Building effective deep neural network architectures one feature at a time

Successful training of convolutional neural networks is often associated...
research
12/11/2021

CHAMP: Coherent Hardware-Aware Magnitude Pruning of Integrated Photonic Neural Networks

We propose a novel hardware-aware magnitude pruning technique for cohere...

Please sign up or login with your details

Forgot password? Click here to reset