Partitioning and Deployment of Deep Neural Networks on Edge Clusters

04/24/2023
by   Arjun Parthasarathy, et al.
0

Edge inference has become more widespread, as its diverse applications range from retail to wearable technology. Clusters of networked resource-constrained edge devices are becoming common, yet no system exists to split a DNN across these clusters while maximizing the inference throughput of the system. Additionally, no production-ready orchestration system exists for deploying said models over such edge networks which adopts the robustness and scalability of the cloud. We present an algorithm which partitions DNNs and distributes them across a set of edge devices with the goal of minimizing the bottleneck latency and therefore maximizing inference throughput. The system scales well to systems of different node memory capacities and numbers of nodes, while being node fault-tolerant. We find that we can reduce the bottleneck latency by 10x over a random algorithm and 35 algorithm, although the joint-partitioning algorithm outperforms our algorithm in most practical use-cases. Furthermore we find empirically that for the set of representative models we tested, the algorithm produces results within 9.2 of the optimal bottleneck latency. We then developed a standalone cluster network emulator on which we tested configurations of up to 20 nodes and found a steady increase in throughput and decrease in end-to-end latency as the cluster size scales. In these tests, we observed that our system has multi-node fault-tolerance as well as network and system IO fault-tolerance. We have implemented our framework in open-source software that is publicly available to the research community at https://github.com/ANRGUSC/SEIFER.

READ FULL TEXT
research
10/21/2022

Partitioning and Placement of Deep Neural Networks on Distributed Edge Devices to Maximize Inference Throughput

Edge inference has become more widespread, as its diverse applications r...
research
10/21/2022

SEIFER: Scalable Edge Inference for Deep Neural Networks

Edge inference is becoming ever prevalent through its applications from ...
research
04/08/2021

Read-Write Quorum Systems Made Practical

Quorum systems are a powerful mechanism for ensuring the consistency of ...
research
01/18/2022

DEFER: Distributed Edge Inference for Deep Neural Networks

Modern machine learning tools such as deep neural networks (DNNs) are pl...
research
09/06/2017

Distributed Deep Neural Networks over the Cloud, the Edge and End Devices

We propose distributed deep neural networks (DDNNs) over distributed com...
research
06/16/2022

Fault-Tolerant Collaborative Inference through the Edge-PRUNE Framework

Collaborative inference has received significant research interest in ma...
research
07/31/2023

DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms

Datacenters are increasingly becoming heterogeneous, and are starting to...

Please sign up or login with your details

Forgot password? Click here to reset