Online Learning for Orchestration of Inference in Multi-User End-Edge-Cloud Networks

02/21/2022
by   Sina Shahhosseini, et al.
5

Deep-learning-based intelligent services have become prevalent in cyber-physical applications including smart cities and health-care. Deploying deep-learning-based intelligence near the end-user enhances privacy protection, responsiveness, and reliability. Resource-constrained end-devices must be carefully managed in order to meet the latency and energy requirements of computationally-intensive deep learning services. Collaborative end-edge-cloud computing for deep learning provides a range of performance and efficiency that can address application requirements through computation offloading. The decision to offload computation is a communication-computation co-optimization problem that varies with both system parameters (e.g., network condition) and workload characteristics (e.g., inputs). On the other hand, deep learning model optimization provides another source of tradeoff between latency and model accuracy. An end-to-end decision-making solution that considers such computation-communication problem is required to synergistically find the optimal offloading policy and model for deep learning services. To this end, we propose a reinforcement-learning-based computation offloading solution that learns optimal offloading policy considering deep learning model selection techniques to minimize response time while providing sufficient accuracy. We demonstrate the effectiveness of our solution for edge devices in an end-edge-cloud system and evaluate with a real-setup implementation using multiple AWS and ARM core configurations. Our solution provides 35 the average response time compared to the state-of-the-art with less than 0.9 accuracy reduction, demonstrating the promise of our online learning framework for orchestrating DL inference in end-edge-cloud systems.

READ FULL TEXT
research
02/21/2022

Hybrid Learning for Orchestrating Deep Learning Inference in Multi-user Edge-cloud Networks

Deep-learning-based intelligent services have become prevalent in cyber-...
research
08/23/2020

DMRO:A Deep Meta Reinforcement Learning-based Task Offloading Framework for Edge-Cloud Computing

With the continuous growth of mobile data and the unprecedented demand f...
research
10/24/2022

Graph Reinforcement Learning-based CNN Inference Offloading in Dynamic Edge Computing

This paper studies the computational offloading of CNN inference in dyna...
research
02/02/2021

Autodidactic Neurosurgeon: Collaborative Deep Inference for Mobile Edge Intelligence via Online Learning

Recent breakthroughs in deep learning (DL) have led to the emergence of ...
research
11/17/2020

Optimal Accuracy-Time Trade-off for Deep Learning Services in Edge Computing Systems

With the increasing demand for computationally intensive services like d...
research
07/03/2019

A Unified Optimization Approach for CNN Model Inference on Integrated GPUs

Modern deep learning applications urge to push the model inference takin...
research
05/06/2020

AutoScale: Optimizing Energy Efficiency of End-to-End Edge Inference under Stochastic Variance

Deep learning inference is increasingly run at the edge. As the programm...

Please sign up or login with your details

Forgot password? Click here to reset