Latency-Memory Optimized Splitting of Convolution Neural Networks for Resource Constrained Edge Devices

07/19/2021
by   Tanmay Jain, et al.
0

With the increasing reliance of users on smart devices, bringing essential computation at the edge has become a crucial requirement for any type of business. Many such computations utilize Convolution Neural Networks (CNNs) to perform AI tasks, having high resource and computation requirements, that are infeasible for edge devices. Splitting the CNN architecture to perform part of the computation on edge and remaining on the cloud is an area of research that has seen increasing interest in the field. In this paper, we assert that running CNNs between an edge device and the cloud is synonymous to solving a resource-constrained optimization problem that minimizes the latency and maximizes resource utilization at the edge. We formulate a multi-objective optimization problem and propose the LMOS algorithm to achieve a Pareto efficient solution. Experiments done on real-world edge devices show that, LMOS ensures feasible execution of different CNN models at the edge and also improves upon existing state-of-the-art approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/01/2021

SmartSplit: Latency-Energy-Memory Optimisation for CNN Splitting on Smartphone Environment

Artificial Intelligence has now taken centre stage in the smartphone ind...
research
09/01/2021

Architecture Aware Latency Constrained Sparse Neural Networks

Acceleration of deep neural networks to meet a specific latency constrai...
research
07/20/2022

AutoDiCE: Fully Automated Distributed CNN Inference at the Edge

Deep Learning approaches based on Convolutional Neural Networks (CNNs) a...
research
10/08/2019

ABEONA: an Architecture for Energy-Aware Task Migrations from the Edge to the Cloud

This paper presents our preliminary results with ABEONA, an edge-to-clou...
research
10/10/2022

DeepPerform: An Efficient Approach for Performance Testing of Resource-Constrained Neural Networks

Today, an increasing number of Adaptive Deep Neural Networks (AdNNs) are...
research
05/14/2020

PENNI: Pruned Kernel Sharing for Efficient CNN Inference

Although state-of-the-art (SOTA) CNNs achieve outstanding performance on...
research
05/21/2022

SplitPlace: AI Augmented Splitting and Placement of Large-Scale Neural Networks in Mobile Edge Environments

In recent years, deep learning models have become ubiquitous in industry...

Please sign up or login with your details

Forgot password? Click here to reset