Training Quantized Nets: A Deeper Understanding

06/07/2017
by   Hao Li, et al.
0

Currently, deep neural networks are deployed on low-power portable devices by first training a full-precision model using powerful hardware, and then deriving a corresponding low-precision model for efficient inference on such systems. However, training models directly with coarsely quantized weights is a key step towards learning on embedded platforms that have limited computing resources, memory capacity, and power consumption. Numerous recent publications have studied methods for training quantized networks, but these studies have mostly been empirical. In this work, we investigate training methods for quantized neural networks from a theoretical viewpoint. We first explore accuracy guarantees for training methods under convexity assumptions. We then look at the behavior of these algorithms for non-convex problems, and show that training algorithms that exploit high-precision representations have an important greedy search phase that purely quantized training methods lack, which explains the difficulty of training using low-precision arithmetic.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2020

Batch Normalization in Quantized Networks

Implementation of quantized neural networks on computing hardware leads ...
research
12/29/2019

MTJ-Based Hardware Synapse Design for Quantized Deep Neural Networks

Quantized neural networks (QNNs) are being actively researched as a solu...
research
02/27/2021

Efficient Soft-Error Detection for Low-precision Deep Learning Recommendation Models

Soft error, namely silent corruption of signal or datum in a computer sy...
research
11/30/2019

Quantized deep learning models on low-power edge devices for robotic systems

In this work, we present a quantized deep neural network deployed on a l...
research
02/02/2021

Benchmarking Quantized Neural Networks on FPGAs with FINN

The ever-growing cost of both training and inference for state-of-the-ar...
research
06/04/2022

Combinatorial optimization for low bit-width neural networks

Low-bit width neural networks have been extensively explored for deploym...
research
05/13/2021

Stochastic-Shield: A Probabilistic Approach Towards Training-Free Adversarial Defense in Quantized CNNs

Quantized neural networks (NN) are the common standard to efficiently de...

Please sign up or login with your details

Forgot password? Click here to reset