SplitEE: Early Exit in Deep Neural Networks with Split Computing

09/17/2023
by   Divya J. Bajpai, et al.
0

Deep Neural Networks (DNNs) have drawn attention because of their outstanding performance on various tasks. However, deploying full-fledged DNNs in resource-constrained devices (edge, mobile, IoT) is difficult due to their large size. To overcome the issue, various approaches are considered, like offloading part of the computation to the cloud for final inference (split computing) or performing the inference at an intermediary layer without passing through all layers (early exits). In this work, we propose combining both approaches by using early exits in split computing. In our approach, we decide up to what depth of DNNs computation to perform on the device (splitting layer) and whether a sample can exit from this layer or need to be offloaded. The decisions are based on a weighted combination of accuracy, computational, and communication costs. We develop an algorithm named SplitEE to learn an optimal policy. Since pre-trained DNNs are often deployed in new domains where the ground truths may be unavailable and samples arrive in a streaming fashion, SplitEE works in an online and unsupervised setup. We extensively perform experiments on five different datasets. SplitEE achieves a significant cost reduction (>50%) with a slight drop in accuracy (<2%) as compared to the case when all samples are inferred at the final layer. The anonymized source code is available at <https://anonymous.4open.science/r/SplitEE_M-B989/README.md>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2022

Dynamic Split Computing for Efficient Deep Edge Intelligence

Deploying deep neural networks (DNNs) on IoT and mobile devices is a cha...
research
10/30/2020

Calibration-Aided Edge Inference Offloading via Adaptive Model Partitioning of Deep Neural Networks

Mobile devices can offload deep neural network (DNN)-based inference to ...
research
09/20/2022

Unsupervised Early Exit in DNNs with Multiple Exits

Deep Neural Networks (DNNs) are generally designed as sequentially casca...
research
09/23/2022

I-SPLIT: Deep Network Interpretability for Split Computing

This work makes a substantial step in the field of split computing, i.e....
research
05/21/2022

SplitPlace: AI Augmented Splitting and Placement of Large-Scale Neural Networks in Mobile Edge Environments

In recent years, deep learning models have become ubiquitous in industry...
research
08/20/2021

Early-exit deep neural networks for distorted images: providing an efficient edge offloading

Edge offloading for deep neural networks (DNNs) can be adaptive to the i...
research
11/05/2018

Efficient Inference on Deep Neural Networks by Dynamic Representations and Decision Gates

The current trade-off between depth and computational cost makes it diff...

Please sign up or login with your details

Forgot password? Click here to reset