SPINN: Synergistic Progressive Inference of Neural Networks over Device and Cloud

08/14/2020
by   Stefanos Laskaridis, et al.
9

Despite the soaring use of convolutional neural networks (CNNs) in mobile applications, uniformly sustaining high-performance inference on mobile has been elusive due to the excessive computational demands of modern CNNs and the increasing diversity of deployed devices. A popular alternative comprises offloading CNN processing to powerful cloud-based servers. Nevertheless, by relying on the cloud to produce outputs, emerging mission-critical and high-mobility applications, such as drone obstacle avoidance or interactive applications, can suffer from the dynamic connectivity conditions and the uncertain availability of the cloud. In this paper, we propose SPINN, a distributed inference system that employs synergistic device-cloud computation together with a progressive inference method to deliver fast and robust CNN inference across diverse settings. The proposed system introduces a novel scheduler that co-optimises the early-exit policy and the CNN splitting at run time, in order to adapt to dynamic conditions and meet user-defined service-level requirements. Quantitative evaluation illustrates that SPINN outperforms its state-of-the-art collaborative inference counterparts by up to 2x in achieved throughput under varying network conditions, reduces the server cost by up to 6.8x and improves accuracy by 20.7 while providing robust operation under uncertain connectivity conditions and significant energy savings compared to cloud-centric execution.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2021

DynO: Dynamic Onloading of Deep Neural Networks from Cloud to Device

Recently, there has been an explosive growth of mobile and embedded appl...
research
08/10/2020

HAPI: Hardware-Aware Progressive Inference

Convolutional neural networks (CNNs) have recently become the state-of-t...
research
10/30/2020

Calibration-Aided Edge Inference Offloading via Adaptive Model Partitioning of Deep Neural Networks

Mobile devices can offload deep neural network (DNN)-based inference to ...
research
07/14/2017

Cloud-based or On-device: An Empirical Study of Mobile Deep Inference

Modern mobile applications are benefiting significantly from the advance...
research
01/25/2018

JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services

Deep neural networks are among the most influential architectures of dee...
research
11/12/2022

PriMask: Cascadable and Collusion-Resilient Data Masking for Mobile Cloud Inference

Mobile cloud offloading is indispensable for inference tasks based on la...
research
11/01/2021

SmartSplit: Latency-Energy-Memory Optimisation for CNN Splitting on Smartphone Environment

Artificial Intelligence has now taken centre stage in the smartphone ind...

Please sign up or login with your details

Forgot password? Click here to reset