DNNAbacus: Toward Accurate Computational Cost Prediction for Deep Neural Networks

05/24/2022
by   Lu Bai, et al.
23

Deep learning is attracting interest across a variety of domains, including natural language processing, speech recognition, and computer vision. However, model training is time-consuming and requires huge computational resources. Existing works on the performance prediction of deep neural networks, which mostly focus on the training time prediction of a few models, rely on analytical models and result in high relative errors. scheduling and reducing job failures in data centers are essential to improve resource utilization and reduce carbon emissions. This paper investigates the computational resource demands of 29 classical deep neural networks and builds accurate models for predicting computational costs. We first analyze the profiling results of typical networks and demonstrate that the computational resource demands of models with different inputs and hyperparameters are not obvious and intuitive. We then propose a lightweight prediction approach DNNAbacus with a novel network structural matrix for network representation. DNNAbacus can accurately predict both memory and time cost for PyTorch and TensorFlow models, which is also generalized to different hardware architectures and can have zero-shot capability for unseen networks. Our experimental results show that the mean relative error (MRE) is 0.9 respect to time and 2.8 much lower than the state-of-the-art works.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2018

Generalization Error in Deep Learning

Deep learning models have lately shown great performance in various fiel...
research
06/20/2023

Deep Fusion: Efficient Network Training via Pre-trained Initializations

In recent years, deep learning has made remarkable progress in a wide ra...
research
01/28/2022

Benchmarking Resource Usage for Efficient Distributed Deep Learning

Deep learning (DL) workflows demand an ever-increasing budget of compute...
research
07/23/2018

NullaNet: Training Deep Neural Networks for Reduced-Memory-Access Inference

Deep neural networks have been successfully deployed in a wide variety o...
research
12/02/2018

Neural Rejuvenation: Improving Deep Network Training by Enhancing Computational Resource Utilization

In this paper, we study the problem of improving computational resource ...
research
12/05/2020

Weight Update Skipping: Reducing Training Time for Artificial Neural Networks

Artificial Neural Networks (ANNs) are known as state-of-the-art techniqu...
research
09/16/2021

Disaggregating and Consolidating Network Functionalities with SuperNIC

Resource disaggregation has gained huge popularity in recent years. Exis...

Please sign up or login with your details

Forgot password? Click here to reset