An optimal scheduling architecture for accelerating batch algorithms on Neural Network processor architectures

02/14/2020
by   Phani Kumar Nyshadham, et al.
0

In neural network topologies, algorithms are running on batches of data tensors. The batches of data are typically scheduled onto the computing cores which execute in parallel. For the algorithms running on batches of data, an optimal batch scheduling architecture is very much needed by suitably utilizing hardware resources - thereby resulting in significant reduction training and inference time. In this paper, we propose to accelerate the batch algorithms for neural networks through a scheduling architecture enabling optimal compute power utilization. The proposed optimal scheduling architecture can be built into HW or can be implemented in SW alone which can be leveraged for accelerating batch algorithms. The results demonstrate that the proposed architecture speeds up the batch algorithms compared to the previous solutions. The proposed idea applies to any HPC architecture meant for neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/16/2020

Extending SLURM for Dynamic Resource-Aware Adaptive Batch Scheduling

With the growing constraints on power budget and increasing hardware fai...
research
11/04/2018

Exploring the Relation Between Two Levels of Scheduling Using a Novel Simulation Approach

Modern high performance computing (HPC) systems exhibit a rapid growth i...
research
05/11/2018

Parallelizing Bisection Root-Finding: A Case for Accelerating Serial Algorithms in Multicore Substrates

Multicore architectures dominate today's processor market. Even though t...
research
09/30/2021

Accelerating Fully Connected Neural Network on Optical Network-on-Chip (ONoC)

Fully Connected Neural Network (FCNN) is a class of Artificial Neural Ne...
research
03/10/2021

A Resourceful Coordination Approach for Multilevel Scheduling

HPC users aim to improve their execution times without particular regard...
research
08/20/2023

Eventually-Consistent Federated Scheduling for Data Center Workloads

Data center schedulers operate at unprecedented scales today to accommod...
research
03/15/2023

Gated Compression Layers for Efficient Always-On Models

Mobile and embedded machine learning developers frequently have to compr...

Please sign up or login with your details

Forgot password? Click here to reset