Deep-Edge: An Efficient Framework for Deep Learning Model Update on Heterogeneous Edge

04/13/2020
by   Anirban Bhattacharjee, et al.
0

Deep Learning (DL) model-based AI services are increasingly offered in a variety of predictive analytics services such as computer vision, natural language processing, speech recognition. However, the quality of the DL models can degrade over time due to changes in the input data distribution, thereby requiring periodic model updates. Although cloud data-centers can meet the computational requirements of the resource-intensive and time-consuming model update task, transferring data from the edge devices to the cloud incurs a significant cost in terms of network bandwidth and are prone to data privacy issues. With the advent of GPU-enabled edge devices, the DL model update can be performed at the edge in a distributed manner using multiple connected edge devices. However, efficiently utilizing the edge resources for the model update is a hard problem due to the heterogeneity among the edge devices and the resource interference caused by the co-location of the DL model update task with latency-critical tasks running in the background. To overcome these challenges, we present Deep-Edge, a load- and interference-aware, fault-tolerant resource management framework for performing model update at the edge that uses distributed training. This paper makes the following contributions. First, it provides a unified framework for monitoring, profiling, and deploying the DL model update tasks on heterogeneous edge devices. Second, it presents a scheduler that reduces the total re-training time by appropriately selecting the edge devices and distributing data among them such that no latency-critical applications experience deadline violations. Finally, we present empirical results to validate the efficacy of the framework using a real-world DL model update case-study based on the Caltech dataset and an edge AI cluster testbed.

READ FULL TEXT

page 1

page 5

page 7

research
07/26/2021

AI Multi-Tenancy on Edge: Concurrent Deep Learning Model Executions and Dynamic Model Placements on Edge Devices

Many real-world applications are widely adopting the edge computing para...
research
04/23/2023

The Case for Hierarchical Deep Learning Inference at the Network Edge

Resource-constrained Edge Devices (EDs), e.g., IoT sensors and microcont...
research
03/16/2021

Distributed Deep Learning Using Volunteer Computing-Like Paradigm

Use of Deep Learning (DL) in commercial applications such as image class...
research
04/11/2019

FECBench: A Holistic Interference-aware Approach for Application Performance Modeling

Services hosted in multi-tenant cloud platforms often encounter performa...
research
04/07/2022

Enabling Deep Learning for All-in EDGE paradigm

Deep Learning-based models have been widely investigated, and they have ...
research
08/29/2023

Generative Model for Models: Rapid DNN Customization for Diverse Tasks and Resource Constraints

Unlike cloud-based deep learning models that are often large and uniform...
research
11/29/2021

Privacy-Preserving Serverless Edge Learning with Decentralized Small Data

In the last decade, data-driven algorithms outperformed traditional opti...

Please sign up or login with your details

Forgot password? Click here to reset