MixBin: Towards Budgeted Binarization

11/12/2022
by   Udbhav Bamba, et al.
10

Binarization has proven to be amongst the most effective ways of neural network compression, reducing the FLOPs of the original model by a large extent. However, such levels of compression are often accompanied by a significant drop in the performance. There exist some approaches that reduce this performance drop by facilitating partial binarization of the network, however, a systematic approach to mix binary and full-precision parameters in a single network is still missing. In this paper, we propose a paradigm to perform partial binarization of neural networks in a controlled sense, thereby constructing budgeted binary neural network (B2NN). We present MixBin, an iterative search-based strategy that constructs B2NN through optimized mixing of the binary and full-precision components. MixBin allows to explicitly choose the approximate fraction of the network to be kept as binary, thereby presenting the flexibility to adapt the inference cost at a prescribed budget. We demonstrate through experiments that B2NNs obtained from our MixBin strategy are significantly better than those obtained from random selection of the network layers. To perform partial binarization in an effective manner, it is important that both the full-precision as well as the binary components of the B2NN are appropriately optimized. We also demonstrate that the choice of the activation function can have a significant effect on this process, and to circumvent this issue, we present BinReLU, that can be used as an effective activation function for the full-precision as well as the binary components of any B2NN. Experimental investigations reveal that BinReLU outperforms the other activation functions in all possible scenarios of B2NN: zero-, partial- as well as full binarization. Finally, we demonstrate the efficacy of MixBin on the tasks of classification and object tracking using benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/16/2023

Data-aware customization of activation functions reduces neural network error

Activation functions play critical roles in neural networks, yet current...
research
10/02/2022

Basic Binary Convolution Unit for Binarized Image Restoration Network

Lighter and faster image restoration (IR) models are crucial for the dep...
research
07/02/2023

ENN: A Neural Network with DCT-Adaptive Activation Functions

The expressiveness of neural networks highly depends on the nature of th...
research
10/12/2020

Activation function impact on Sparse Neural Networks

While the concept of a Sparse Neural Network has been researched for som...
research
09/30/2019

Random Bias Initialization Improving Binary Neural Network Training

Edge intelligence especially binary neural network (BNN) has attracted c...
research
05/24/2022

Taming the sign problem of explicitly antisymmetrized neural networks via rough activation functions

Explicit antisymmetrization of a two-layer neural network is a potential...
research
06/18/2020

Short Communication: Detecting Possibly Frequent Change-points: Wild Binary Segmentation 2

This article comments on the new version of wild binary segmentation 2. ...

Please sign up or login with your details

Forgot password? Click here to reset