Neural Architecture Search for Improving Latency-Accuracy Trade-off in Split Computing

08/30/2022
by   Shoma Shimizu, et al.
7

This paper proposes a neural architecture search (NAS) method for split computing. Split computing is an emerging machine-learning inference technique that addresses the privacy and latency challenges of deploying deep learning in IoT systems. In split computing, neural network models are separated and cooperatively processed using edge servers and IoT devices via networks. Thus, the architecture of the neural network model significantly impacts the communication payload size, model accuracy, and computational load. In this paper, we address the challenge of optimizing neural network architecture for split computing. To this end, we proposed NASC, which jointly explores optimal model architecture and a split point to achieve higher accuracy while meeting latency requirements (i.e., smaller total latency of computation and communication than a certain threshold). NASC employs a one-shot NAS that does not require repeating model training for a computationally efficient architecture search. Our performance evaluation using hardware (HW)-NAS-Bench of benchmark data demonstrates that the proposed NASC can improve the “communication latency and model accuracy" trade-off, i.e., reduce the latency by approximately 40-60

READ FULL TEXT
research
11/16/2021

JMSNAS: Joint Model Split and Neural Architecture Search for Learning over Mobile Edge Networks

The main challenge to deploy deep neural network (DNN) over a mobile edg...
research
11/03/2021

Communication-Efficient Separable Neural Network for Distributed Inference on Edge Devices

The inference of Neural Networks is usually restricted by the resources ...
research
04/25/2021

Balancing Accuracy and Latency in Multipath Neural Networks

The growing capacity of neural networks has strongly contributed to thei...
research
06/21/2021

ESAI: Efficient Split Artificial Intelligence via Early Exiting Using Neural Architecture Search

Recently, deep neural networks have been outperforming conventional mach...
research
10/21/2020

MicroNets: Neural Network Architectures for Deploying TinyML Applications on Commodity Microcontrollers

Executing machine learning workloads locally on resource constrained mic...
research
08/13/2021

Spatio-Temporal Split Learning

This paper proposes a novel split learning framework with multiple end-s...
research
09/06/2019

Distributed creation of Machine learning agents for Blockchain analysis

Creating efficient deep neural networks involves repetitive manual optim...

Please sign up or login with your details

Forgot password? Click here to reset