Learning Pruned Structure and Weights Simultaneously from Scratch: an Attention based Approach

11/01/2021
by   Qisheng He, et al.
0

As a deep learning model typically contains millions of trainable weights, there has been a growing demand for a more efficient network structure with reduced storage space and improved run-time efficiency. Pruning is one of the most popular network compression techniques. In this paper, we propose a novel unstructured pruning pipeline, Attention-based Simultaneous sparse structure and Weight Learning (ASWL). Unlike traditional channel-wise or weight-wise attention mechanism, ASWL proposed an efficient algorithm to calculate the pruning ratio through layer-wise attention for each layer, and both weights for the dense network and the sparse network are tracked so that the pruned structure is simultaneously learned from randomly initialized weights. Our experiments on MNIST, Cifar10, and ImageNet show that ASWL achieves superior pruning results in terms of accuracy, pruning ratio and operating efficiency when compared with state-of-the-art network pruning methods.

READ FULL TEXT

page 2

page 7

research
06/08/2023

Magnitude Attention-based Dynamic Pruning

Existing pruning methods utilize the importance of each weight based on ...
research
10/08/2021

ABCP: Automatic Block-wise and Channel-wise Network Pruning via Joint Search

Currently, an increasing number of model pruning methods are proposed to...
research
05/14/2020

Dynamic Sparse Training: Find Efficient Sparse Network From Scratch With Trainable Masked Layers

We present a novel network pruning algorithm called Dynamic Sparse Train...
research
10/24/2022

Weight Fixing Networks

Modern iterations of deep learning models contain millions (billions) of...
research
01/31/2022

SPDY: Accurate Pruning with Speedup Guarantees

The recent focus on the efficiency of deep neural networks (DNNs) has le...
research
03/28/2023

Learning Second-Order Attentive Context for Efficient Correspondence Pruning

Correspondence pruning aims to search consistent correspondences (inlier...
research
07/04/2020

Weight-dependent Gates for Network Pruning

In this paper, we propose a simple and effective network pruning framewo...

Please sign up or login with your details

Forgot password? Click here to reset