Compute, Time and Energy Characterization of Encoder-Decoder Networks with Automatic Mixed Precision Training

08/18/2020
by   Siddharth Samsi, et al.
0

Deep neural networks have shown great success in many diverse fields. The training of these networks can take significant amounts of time, compute and energy. As datasets get larger and models become more complex, the exploration of model architectures becomes prohibitive. In this paper we examine the compute, energy and time costs of training a UNet based deep neural network for the problem of predicting short term weather forecasts (called precipitation Nowcasting). By leveraging a combination of data distributed and mixed-precision training, we explore the design space for this problem. We also show that larger models with better performance come at a potentially incremental cost if appropriate optimizations are used. We show that it is possible to achieve a significant improvement in training time by leveraging mixed-precision training without sacrificing model performance. Additionally, we find that a 1549 network comes at a relatively smaller 63.22 UNet with 4 encoding layers.

READ FULL TEXT
research
08/28/2019

Distributed Deep Learning for Precipitation Nowcasting

Effective training of Deep Neural Networks requires massive amounts of d...
research
01/30/2023

Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference

For effective and efficient deep neural network inference, it is desirab...
research
07/23/2015

Neural NILM: Deep Neural Networks Applied to Energy Disaggregation

Energy disaggregation estimates appliance-by-appliance electricity consu...
research
10/28/2019

Adaptive Loss Scaling for Mixed Precision Training

Mixed precision training (MPT) is becoming a practical technique to impr...
research
07/10/2023

QBitOpt: Fast and Accurate Bitwidth Reallocation during Training

Quantizing neural networks is one of the most effective methods for achi...
research
07/27/2023

Speeding up Fourier Neural Operators via Mixed Precision

The Fourier neural operator (FNO) is a powerful technique for learning s...
research
02/15/2022

Predicting on the Edge: Identifying Where a Larger Model Does Better

Much effort has been devoted to making large and more accurate models, b...

Please sign up or login with your details

Forgot password? Click here to reset