Dynamic Hard Pruning of Neural Networks at the Edge of the Internet

11/17/2020
by   Lorenzo Valerio, et al.
0

Neural Networks (NN), although successfully applied to several Artificial Intelligence tasks, are often unnecessarily over-parametrized. In fog/edge computing, this might make their training prohibitive on resource-constrained devices, contrasting with the current trend of decentralising intelligence from remote data-centres to local constrained devices. Therefore, we investigate the problem of training effective NN models on constrained devices having a fixed, potentially small, memory budget. We target techniques that are both resource-efficient and performance effective while enabling significant network compression. Our technique, called Dynamic Hard Pruning (DynHP), incrementally prunes the network during training, identifying neurons that marginally contribute to the model accuracy. DynHP enables a tunable size reduction of the final neural network and reduces the NN memory occupancy during training. Freed memory is reused by a dynamic batch sizing approach to counterbalance the accuracy degradation caused by the hard pruning strategy, improving its convergence and effectiveness. We assess the performance of DynHP through reproducible experiments on two public datasets, comparing them against reference competitors. Results show that DynHP compresses a NN up to 10 times without significant performance drops (up to 5% relative error w.r.t. competitors), reducing up to 80% the training memory occupancy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/16/2021

Successive Pruning for Model Compression via Rate Distortion Theory

Neural network (NN) compression has become essential to enable deploying...
research
03/05/2020

Pruning Filters while Training for Efficiently Optimizing Deep Learning Networks

Modern deep networks have millions to billions of parameters, which lead...
research
01/14/2021

Enabling Large Neural Networks on Tiny Microcontrollers with Swapping

Running neural networks (NNs) on microcontroller units (MCUs) is becomin...
research
10/27/2021

Binarized ResNet: Enabling Automatic Modulation Classification at the resource-constrained Edge

In this paper, we propose a ResNet based neural architecture to solve th...
research
05/30/2019

Toward Runtime-Throttleable Neural Networks

As deep neural network (NN) methods have matured, there has been increas...
research
01/22/2021

Baseline Pruning-Based Approach to Trojan Detection in Neural Networks

This paper addresses the problem of detecting trojans in neural networks...
research
06/22/2022

GACT: Activation Compressed Training for General Architectures

Training large neural network (NN) models requires extensive memory reso...

Please sign up or login with your details

Forgot password? Click here to reset