An End-to-End HW/SW Co-Design Methodology to Design Efficient Deep Neural Network Systems using Virtual Models

10/25/2019
by   Michael J. Klaiber, et al.
0

End-to-end performance estimation and measurement of deep neural network (DNN) systems become more important with increasing complexity of DNN systems consisting of hardware and software components. The methodology proposed in this paper aims at a reduced turn-around time for evaluating different design choices of hardware and software components of DNN systems. This reduction is achieved by moving the performance estimation from the implementation phase to the concept phase by employing virtual hardware models instead of gathering measurement results from physical prototypes. Deep learning compilers introduce hardware-specific transformations and are, therefore, considered a part of the design flow of virtual system models to extract end-to-end performance estimations. To validate the run-time accuracy of the proposed methodology, a system processing the DilatedVGG DNN is realized both as virtual system model and as hardware implementation. The results show that up to 92 be reached in predicting the processing time of the DNN inference.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/31/2023

Fused Depthwise Tiling for Memory Optimization in TinyML Deep Neural Network Inference

Memory optimization for deep neural network (DNN) inference gains high r...
research
09/04/2020

CLEANN: Accelerated Trojan Shield for Embedded Neural Networks

We propose CLEANN, the first end-to-end framework that enables online mi...
research
12/28/2020

SimBricks: End-to-End Network System Evaluation with Modular Simulation

Full system "end-to-end" measurements in physical testbeds are the gold ...
research
09/20/2020

VirtualFlow: Decoupling Deep Learning Model Execution from Underlying Hardware

State-of-the-art deep learning systems tightly couple model execution wi...
research
10/18/2022

Generalized Many-Body Dispersion Correction through Random-phase Approximation for Chemically Accurate Density Functional Theory

We extend our recently proposed Deep Learning-aided many-body dispersion...
research
11/03/2017

SparseNN: An Energy-Efficient Neural Network Accelerator Exploiting Input and Output Sparsity

Contemporary Deep Neural Network (DNN) contains millions of synaptic con...

Please sign up or login with your details

Forgot password? Click here to reset