A Greedy Algorithm for Building Compact Binary Activated Neural Networks

09/07/2022
by   Benjamin Leblanc, et al.
0

We study binary activated neural networks in the context of regression tasks, provide guarantees on the expressiveness of these particular networks and propose a greedy algorithm for building such networks. Aiming for predictors having small resources needs, the greedy approach does not need to fix in advance an architecture for the network: this one is built one layer at a time, one neuron at a time, leading to predictors that aren't needlessly wide and deep for a given task. Similarly to boosting algorithms, our approach guarantees a training loss reduction every time a neuron is added to a layer. This greatly differs from most binary activated neural networks training schemes that rely on stochastic gradient descent (circumventing the 0-almost-everywhere derivative problem of the binary activation function by surrogates such as the straight through estimator or continuous binarization). We show that our method provides compact and sparse predictors while obtaining similar performances to state-of-the-art methods for training binary activated networks.

READ FULL TEXT
research
10/14/2018

Variational Neural Networks: Every Layer and Neuron Can Be Unique

The choice of activation function can significantly influence the perfor...
research
04/06/2023

Training a Two Layer ReLU Network Analytically

Neural networks are usually trained with different variants of gradient ...
research
10/28/2019

Growing axons: greedy learning of neural networks with application to function approximation

We propose a new method for learning deep neural network models that is ...
research
04/05/2020

On Tractable Representations of Binary Neural Networks

We consider the compilation of a binary neural network's decision functi...
research
02/23/2019

On Greedy Algorithms for Binary de Bruijn Sequences

We propose a general greedy algorithm for binary de Bruijn sequences, ca...
research
04/24/2020

Nonconvex penalization for sparse neural networks

Training methods for artificial neural networks often rely on over-param...
research
10/11/2015

Neural Networks with Few Multiplications

For most deep learning algorithms training is notoriously time consuming...

Please sign up or login with your details

Forgot password? Click here to reset