Stochastic Markov Gradient Descent and Training Low-Bit Neural Networks

08/25/2020
by   Jonathan Ashbrock, et al.
0

The massive size of modern neural networks has motivated substantial recent interest in neural network quantization. We introduce Stochastic Markov Gradient Descent (SMGD), a discrete optimization method applicable to training quantized neural networks. The SMGD algorithm is designed for settings where memory is highly constrained during training. We provide theoretical guarantees of algorithm performance as well as encouraging numerical results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/22/2022

Langevin algorithms for Markovian Neural Networks and Deep Stochastic control

Stochastic Gradient Descent Langevin Dynamics (SGLD) algorithms, which a...
research
08/15/2018

Blended Coarse Gradient Descent for Full Quantization of Deep Neural Networks

Quantized deep neural networks (QDNNs) are attractive due to their much ...
research
12/11/2018

Proximal Mean-field for Neural Network Quantization

Compressing large neural networks by quantizing the parameters, while ma...
research
10/18/2019

Mirror Descent View for Neural Network Quantization

Quantizing large Neural Networks (NN) while maintaining the performance ...
research
01/16/2013

Big Neural Networks Waste Capacity

This article exposes the failure of some big neural networks to leverage...
research
10/03/2022

Limitations of neural network training due to numerical instability of backpropagation

We study the training of deep neural networks by gradient descent where ...
research
12/14/2017

Nonparametric Neural Networks

Automatically determining the optimal size of a neural network for a giv...

Please sign up or login with your details

Forgot password? Click here to reset