All-in-One: A Highly Representative DNN Pruning Framework for Edge Devices with Dynamic Power Management

12/09/2022
by   Yifan Gong, et al.
0

During the deployment of deep neural networks (DNNs) on edge devices, many research efforts are devoted to the limited hardware resource. However, little attention is paid to the influence of dynamic power management. As edge devices typically only have a budget of energy with batteries (rather than almost unlimited energy support on servers or workstations), their dynamic power management often changes the execution frequency as in the widely-used dynamic voltage and frequency scaling (DVFS) technique. This leads to highly unstable inference speed performance, especially for computation-intensive DNN models, which can harm user experience and waste hardware resources. We firstly identify this problem and then propose All-in-One, a highly representative pruning framework to work with dynamic power management using DVFS. The framework can use only one set of model weights and soft masks (together with other auxiliary parameters of negligible storage) to represent multiple models of various pruning ratios. By re-configuring the model to the corresponding pruning ratio for a specific execution frequency (and voltage), we are able to achieve stable inference speed, i.e., keeping the difference in speed performance under various execution frequencies as small as possible. Our experiments demonstrate that our method not only achieves high accuracy for multiple models of different pruning ratios, but also reduces their variance of inference latency for various frequencies, with minimal memory consumption of only one model and one soft mask.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2023

DVFO: Learning-Based DVFS for Energy-Efficient Edge-Cloud Collaborative Inference

Due to limited resources on edge and different characteristics of deep n...
research
06/13/2021

Adaptive Dynamic Pruning for Non-IID Federated Learning

Federated Learning (FL) has emerged as a new paradigm of training machin...
research
11/08/2020

Channel Pruning Guided by Spatial and Channel Attention for DNNs in Intelligent Edge Computing

Deep Neural Networks (DNNs) have achieved remarkable success in many com...
research
07/17/2023

Differentiable Transportation Pruning

Deep learning algorithms are increasingly employed at the edge. However,...
research
05/08/2021

Incremental Training and Group Convolution Pruning for Runtime DNN Performance Scaling on Heterogeneous Embedded Platforms

Inference for Deep Neural Networks is increasingly being executed locall...
research
06/09/2020

Automated Design Space Exploration for optimised Deployment of DNN on Arm Cortex-A CPUs

The spread of deep learning on embedded devices has prompted the develop...
research
05/27/2019

Efficient Network Construction through Structural Plasticity

Deep Neural Networks (DNNs) on hardware is facing excessive computation ...

Please sign up or login with your details

Forgot password? Click here to reset