Evolutionary Training of Sparse Artificial Neural Networks: A Network Science Perspective

Through the success of deep learning, Artificial Neural Networks (ANNs) are among the most used artificial intelligence methods nowadays. ANNs have led to major breakthroughs in various domains, such as particle physics, reinforcement learning, speech recognition, computer vision, and so on. Taking inspiration from the network properties of biological neural networks (e.g. sparsity, scale-freeness), we argue that (contrary to general practice) Artificial Neural Networks (ANN), too, should not have fully-connected layers. We show how ANNs perform perfectly well with sparsely-connected layers. Following a Darwinian evolutionary approach, we propose a novel algorithm which evolves an initial random sparse topology (i.e. an Erdős-Rényi random graph) of two consecutive layers of neurons into a scale-free topology, during the ANN training process. The resulting sparse layers can safely replace the corresponding fully-connected layers. Our method allows to quadratically reduce the number of parameters in the fully conencted layers of ANNs, yielding quadratically faster computational times in both phases (i.e. training and inference), at no decrease in accuracy. We demonstrate our claims on two popular ANN types (restricted Boltzmann machine and multi-layer perceptron), on two types of tasks (supervised and unsupervised learning), and on 14 benchmark datasets. We anticipate that our approach will enable ANNs having billions of neurons and evolved topologies to be capable of handling complex real-world tasks that are intractable using state-of-the-art methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/17/2023

Connected Hidden Neurons (CHNNet): An Artificial Neural Network for Rapid Convergence

The core purpose of developing artificial neural networks was to mimic t...
research
03/13/2021

Efficient Sparse Artificial Neural Networks

The brain, as the source of inspiration for Artificial Neural Networks (...
research
04/20/2016

A topological insight into restricted Boltzmann machines

Restricted Boltzmann Machines (RBMs) and models derived from them have b...
research
04/16/2021

ALF – A Fitness-Based Artificial Life Form for Evolving Large-Scale Neural Networks

Machine Learning (ML) is becoming increasingly important in daily life. ...
research
07/29/2021

Structure and Performance of Fully Connected Neural Networks: Emerging Complex Network Properties

Understanding the behavior of Artificial Neural Networks is one of the m...
research
02/12/2021

Exploiting Spline Models for the Training of Fully Connected Layers in Neural Network

The fully connected (FC) layer, one of the most fundamental modules in a...
research
06/09/2023

Weight Freezing: A Regularization Approach for Fully Connected Layers with an Application in EEG Classification

In the realm of EEG decoding, enhancing the performance of artificial ne...

Please sign up or login with your details

Forgot password? Click here to reset