Neural Compression and Filtering for Edge-assisted Real-time Object Detection in Challenged Networks

by   Yoshitomo Matsubara, et al.

The edge computing paradigm places compute-capable devices - edge servers - at the network edge to assist mobile devices in executing data analysis tasks. Intuitively, offloading compute-intense tasks to edge servers can reduce their execution time. However, poor conditions of the wireless channel connecting the mobile devices to the edge servers may degrade the overall capture-to-output delay achieved by edge offloading. Herein, we focus on edge computing supporting remote object detection by means of Deep Neural Networks (DNNs), and develop a framework to reduce the amount of data transmitted over the wireless link. The core idea we propose builds on recent approaches splitting DNNs into sections - namely head and tail models - executed by the mobile device and edge server, respectively. The wireless link, then, is used to transport the output of the last layer of the head model to the edge server, instead of the DNN input. Most prior work focuses on classification tasks and leaves the DNN structure unaltered. Herein, our focus is on DNNs for three different object detection tasks, which present a much more convoluted structure, and modify the architecture of the network to: (i) achieve in-network compression by introducing a bottleneck layer in the early layers on the head model, and (ii) prefilter pictures that do not contain objects of interest using a convolutional neural network. Results show that the proposed technique represents an effective intermediate option between local and edge computing in a parameter region where these extreme point solutions fail to provide satisfactory performance. We release the code and trained models at .


page 1

page 6

page 10


Improving Device-Edge Cooperative Inference of Deep Learning via 2-Step Pruning

Deep neural networks (DNNs) are state-of-the-art solutions for many mach...

Split Computing for Complex Object Detectors: Challenges and Preliminary Results

Following the trends of mobile and edge computing for DNN models, an int...

Slimmable Encoders for Flexible Split DNNs in Bandwidth and Resource Constrained IoT Systems

The execution of large deep neural networks (DNN) at mobile edge devices...

FrankenSplit: Saliency Guided Neural Feature Compression with Shallow Variational Bottleneck Injection

The rise of mobile AI accelerators allows latency-sensitive applications...

Network-Aware 5G Edge Computing for Object Detection: Augmenting Wearables to "See” More, Farther and Faster

Advanced wearable devices are increasingly incorporating high-resolution...

Deep Learning on Mobile Devices Through Neural Processing Units and Edge Computing

Deep Neural Network (DNN) is becoming adopted for video analytics on mob...

A Geo-Aware Server Assignment Problem for Mobile Edge Computing

As mobile devices have become the preferred tool for communication, work...

Please sign up or login with your details

Forgot password? Click here to reset