Training Deep Neural Networks with Constrained Learning Parameters

09/01/2020
by   Prasanna Date, et al.
25

Today's deep learning models are primarily trained on CPUs and GPUs. Although these models tend to have low error, they consume high power and utilize large amount of memory owing to double precision floating point learning parameters. Beyond the Moore's law, a significant portion of deep learning tasks would run on edge computing systems, which will form an indispensable part of the entire computation fabric. Subsequently, training deep learning models for such systems will have to be tailored and adopted to generate models that have the following desirable characteristics: low error, low memory, and low power. We believe that deep neural networks (DNNs), where learning parameters are constrained to have a set of finite discrete values, running on neuromorphic computing systems would be instrumental for intelligent edge computing systems having these desirable characteristics. To this extent, we propose the Combinatorial Neural Network Training Algorithm (CoNNTrA), that leverages a coordinate gradient descent-based approach for training deep learning models with finite discrete learning parameters. Next, we elaborate on the theoretical underpinnings and evaluate the computational complexity of CoNNTrA. As a proof of concept, we use CoNNTrA to train deep learning models with ternary learning parameters on the MNIST, Iris and ImageNet data sets and compare their performance to the same models trained using Backpropagation. We use following performance metrics for the comparison: (i) Training error; (ii) Validation error; (iii) Memory usage; and (iv) Training time. Our results indicate that CoNNTrA models use 32x less memory and have errors at par with the Backpropagation models.

READ FULL TEXT

page 1

page 8

research
12/22/2014

Training deep neural networks with low precision multiplications

Multipliers are the most space and power-hungry arithmetic operators of ...
research
06/16/2020

Gradient Amplification: An efficient way to train deep neural networks

Improving performance of deep learning models and reducing their trainin...
research
10/21/2021

Analysis of memory consumption by neural networks based on hyperparameters

Deep learning models are trained and deployed in multiple domains. Incre...
research
06/15/2022

Edge Inference with Fully Differentiable Quantized Mixed Precision Neural Networks

The large computing and memory cost of deep neural networks (DNNs) often...
research
11/29/2020

Scaling down Deep Learning

Though deep learning models have taken on commercial and political relev...
research
11/08/2016

An Efficient Approach to Boosting Performance of Deep Spiking Network Training

Nowadays deep learning is dominating the field of machine learning with ...
research
04/20/2023

Backpropagation-free Training of Deep Physical Neural Networks

Recent years have witnessed the outstanding success of deep learning in ...

Please sign up or login with your details

Forgot password? Click here to reset