Compact Neural Representation Using Attentive Network Pruning

05/10/2020
by   Mahdi Biparva, et al.
0

Deep neural networks have evolved to become power demanding and consequently difficult to apply to small-size mobile platforms. Network parameter reduction methods have been introduced to systematically deal with the computational and memory complexity of deep networks. We propose to examine the ability of attentive connection pruning to deal with redundancy reduction in neural networks as a contribution to the reduction of computational demand. In this work, we describe a Top-Down attention mechanism that is added to a Bottom-Up feedforward network to select important connections and subsequently prune redundant ones at all parametric layers. Our method not only introduces a novel hierarchical selection mechanism as the basis of pruning but also remains competitive with previous baseline methods in the experimental evaluation. We conduct experiments using different network architectures on popular benchmark datasets to show high compression ratio is achievable with negligible loss of accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/08/2021

Convolutional Neural Network Pruning with Structural Redundancy Reduction

Convolutional neural network (CNN) pruning has become one of the most su...
research
03/19/2021

Toward Compact Deep Neural Networks via Energy-Aware Pruning

Despite of the remarkable performance, modern deep neural networks are i...
research
03/15/2018

Efficient Hardware Realization of Convolutional Neural Networks using Intra-Kernel Regular Pruning

The recent trend toward increasingly deep convolutional neural networks ...
research
06/14/2018

PCAS: Pruning Channels with Attention Statistics

To implement deep neural networks on small embedded devices, conventiona...
research
10/05/2017

To prune, or not to prune: exploring the efficacy of pruning for model compression

Model pruning seeks to induce sparsity in a deep neural network's variou...
research
08/16/2016

Dynamic Network Surgery for Efficient DNNs

Deep learning has become a ubiquitous technology to improve machine inte...
research
11/23/2020

Synthesis and Pruning as a Dynamic Compression Strategy for Efficient Deep Neural Networks

The brain is a highly reconfigurable machine capable of task-specific ad...

Please sign up or login with your details

Forgot password? Click here to reset