Communication-Computation Trade-Off in Resource-Constrained Edge Inference

06/03/2020
by   Jiawei Shao, et al.
0

The recent breakthrough in artificial intelligence (AI), especially deep neural networks (DNNs), has affected every branch of science and technology. Particularly, edge AI has been envisioned as a major application scenario to provide DNN-based services at edge devices. This article presents effective methods for edge inference at resource-constrained devices. It focuses on device-edge co-inference, assisted by an edge computing server, and investigates a critical trade-off among the computation cost of the on-device model and the communication cost of forwarding the intermediate feature to the edge server. A three-step framework is proposed for the effective inference: (1) model split point selection to determine the on-device model, (2) communication-aware model compression to reduce the on-device computation and the resulting communication overhead simultaneously, and (3) task-oriented encoding of the intermediate feature to further reduce the communication overhead. Experiments demonstrate that our proposed framework achieves a better trade-off and significantly reduces the inference latency than baseline methods.

READ FULL TEXT
research
08/30/2021

Communication-Computation Efficient Device-Edge Co-Inference via AutoML

Device-edge co-inference, which partitions a deep neural network between...
research
06/15/2022

Resource-Constrained Edge AI with Early Exit Prediction

By leveraging the data sample diversity, the early-exit network recently...
research
10/27/2020

Branchy-GNN: a Device-Edge Co-Inference Framework for Efficient Point Cloud Processing

The recent advancements of three-dimensional (3D) data acquisition devic...
research
11/15/2022

Enabling AI Quality Control via Feature Hierarchical Edge Inference

With the rise of edge computing, various AI services are expected to be ...
research
12/14/2021

Progressive Feature Transmission for Split Inference at the Wireless Edge

In edge inference, an edge server provides remote-inference services to ...
research
12/10/2021

Towards Homomorphic Inference Beyond the Edge

Beyond edge devices can function off the power grid and without batterie...
research
04/23/2021

Unsupervised Information Obfuscation for Split Inference of Neural Networks

Splitting network computations between the edge device and a server enab...

Please sign up or login with your details

Forgot password? Click here to reset