AdderNet: Do We Really Need Multiplications in Deep Learning?

12/31/2019
by   Hanting Chen, et al.
31

Compared with cheap addition operation, multiplication operation is of much higher computation complexity. The widely-used convolutions in deep neural networks are exactly cross-correlation to measure the similarity between input feature and convolution filters, which involves massive multiplications between float values. In this paper, we present adder networks (AdderNets) to trade these massive multiplications in deep neural networks, especially convolutional neural networks (CNNs), for much cheaper additions to reduce computation costs. In AdderNets, we take the ℓ_1-norm distance between filters and input feature as the output response. The influence of this new similarity measure on the optimization of neural network have been thoroughly analyzed. To achieve a better performance, we develop a special back-propagation approach for AdderNets by investigating the full-precision gradient. We then propose an adaptive learning rate strategy to enhance the training procedure of AdderNets according to the magnitude of each neuron's gradient. As a result, the proposed AdderNets can achieve 74.9 on the ImageNet dataset without any multiplication in convolution layer.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2019

Modulated binary cliquenet

Although Convolutional Neural Networks (CNNs) achieve effectiveness in v...
research
06/05/2022

Searching Similarity Measure for Binarized Neural Networks

Being a promising model to be deployed in resource-limited devices, Bina...
research
02/27/2018

Train Feedfoward Neural Network with Layer-wise Adaptive Rate via Approximating Back-matching Propagation

Stochastic gradient descent (SGD) has achieved great success in training...
research
02/06/2021

IC Networks: Remodeling the Basic Unit for Convolutional Neural Networks

Convolutional neural network (CNN) is a class of artificial neural netwo...
research
05/12/2021

Winograd Algorithm for AdderNet

Adder neural network (AdderNet) is a new kind of deep model that replace...
research
05/30/2019

A Trainable Multiplication Layer for Auto-correlation and Co-occurrence Extraction

In this paper, we propose a trainable multiplication layer (TML) for a n...
research
09/20/2021

Learning Versatile Convolution Filters for Efficient Visual Recognition

This paper introduces versatile filters to construct efficient convoluti...

Please sign up or login with your details

Forgot password? Click here to reset