Fault-Tolerant Collaborative Inference through the Edge-PRUNE Framework

06/16/2022
by   Jani Boutellier, et al.
0

Collaborative inference has received significant research interest in machine learning as a vehicle for distributing computation load, reducing latency, as well as addressing privacy preservation in communications. Recent collaborative inference frameworks have adopted dynamic inference methodologies such as early-exit and run-time partitioning of neural networks. However, as machine learning frameworks scale in the number of inference inputs, e.g., in surveillance applications, fault tolerance related to device failure needs to be considered. This paper presents the Edge-PRUNE distributed computing framework, built on a formally defined model of computation, which provides a flexible infrastructure for fault tolerant collaborative inference. The experimental section of this work shows results on achievable inference time savings by collaborative inference, presents fault tolerant system topologies and analyzes their cost in terms of execution time overhead.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/27/2022

Edge-PRUNE: Flexible Distributed Deep Learning Inference

Collaborative deep learning inference between low-resource endpoint devi...
research
01/11/2021

A Fault Tolerant Mechanism for Partitioning and Offloading Framework in Pervasive Environments

Application partitioning and code offloading are being researched extens...
research
05/01/2023

Anatomy of High-Performance GEMM with Online Fault Tolerance on GPUs

General Matrix Multiplication (GEMM) is a crucial algorithm for various ...
research
03/25/2021

Actuator Fault-Tolerant Vehicle Motion Control: A Survey

The advent of automated vehicles operating at SAE levels 4 and 5 poses h...
research
10/06/2021

FTPipeHD: A Fault-Tolerant Pipeline-Parallel Distributed Training Framework for Heterogeneous Edge Devices

With the increased penetration and proliferation of Internet of Things (...
research
11/24/2022

Design and Prototyping Distributed CNN Inference Acceleration in Edge Computing

For time-critical IoT applications using deep learning, inference accele...
research
04/24/2023

Partitioning and Deployment of Deep Neural Networks on Edge Clusters

Edge inference has become more widespread, as its diverse applications r...

Please sign up or login with your details

Forgot password? Click here to reset