AVEC: Accelerator Virtualization in Cloud-Edge Computing for Deep Learning Libraries

03/08/2021
by   Jason Kennedy, et al.
0

Edge computing offers the distinct advantage of harnessing compute capabilities on resources located at the edge of the network to run workloads of relatively weak user devices. This is achieved by offloading computationally intensive workloads, such as deep learning from user devices to the edge. Using the edge reduces the overall communication latency of applications as workloads can be processed closer to where data is generated on user devices rather than sending them to geographically distant clouds. Specialised hardware accelerators, such as Graphics Processing Units (GPUs) available in the cloud-edge network can enhance the performance of computationally intensive workloads that are offloaded from devices on to the edge. The underlying approach required to facilitate this is virtualization of GPUs. This paper therefore sets out to investigate the potential of GPU accelerator virtualization to improve the performance of deep learning workloads in a cloud-edge environment. The AVEC accelerator virtualization framework is proposed that incurs minimum overheads and requires no source-code modification of the workload. AVEC intercepts local calls to a GPU on a device and forwards them to an edge resource seamlessly. The feasibility of AVEC is demonstrated on a real-world application, namely OpenPose using the Caffe deep learning library. It is observed that on a lab-based experimental test-bed AVEC delivers up to 7.48x speedup despite communication overheads incurred due to data transfers.

READ FULL TEXT

page 1

page 3

research
12/07/2020

Cost-effective Machine Learning Inference Offload for Edge Computing

Computing at the edge is increasingly important since a massive amount o...
research
05/24/2023

Reconfigurable Distributed FPGA Cluster Design for Deep Learning Accelerators

We propose a distributed system based on lowpower embedded FPGAs designe...
research
02/10/2020

AI-oriented Medical Workload Allocation for Hierarchical Cloud/Edge/Device Computing

In a hierarchically-structured cloud/edge/device computing environment, ...
research
10/07/2021

MAPA: Multi-Accelerator Pattern Allocation Policy for Multi-Tenant GPU Servers

Multi-accelerator servers are increasingly being deployed in shared mult...
research
01/12/2023

Reaching the Edge of the Edge: Image Analysis in Space

Satellites have become more widely available due to the reduction in siz...
research
04/10/2023

Deploying Machine Learning Models to Ahead-of-Time Runtime on Edge Using MicroTVM

In the past few years, more and more AI applications have been applied t...
research
05/21/2019

Performance Analysis of Deep Learning Workloads on Leading-edge Systems

This work examines the performance of leading-edge systems designed for ...

Please sign up or login with your details

Forgot password? Click here to reset