SWNet: Small-World Neural Networks and Rapid Convergence

04/09/2019
by   Mojan Javaheripi, et al.
0

Training large and highly accurate deep learning (DL) models is computationally costly. This cost is in great part due to the excessive number of trained parameters, which are well-known to be redundant and compressible for the execution phase. This paper proposes a novel transformation which changes the topology of the DL architecture such that it reaches an optimal cross-layer connectivity. This transformation leverages our important observation that for a set level of accuracy, convergence is fastest when network topology reaches the boundary of a Small-World Network. Small-world graphs are known to possess a specific connectivity structure that enables enhanced signal propagation among nodes. Our small-world models, called SWNets, provide several intriguing benefits: they facilitate data (gradient) flow within the network, enable feature-map reuse by adding long-range connections and accommodate various network architectures/datasets. Compared to densely connected networks (e.g., DenseNets), SWNets require a substantially fewer number of training parameters while maintaining a similar level of classification accuracy. We evaluate our networks on various DL model architectures and image classification datasets, namely, CIFAR10, CIFAR100, and ILSVRC (ImageNet). Our experiments demonstrate an average of 2.1x improvement in convergence speed to the desired accuracy

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2018

Exploring Feature Reuse in DenseNet Architectures

Densely Connected Convolutional Networks (DenseNets) have been shown to ...
research
07/04/2018

Selective Deep Convolutional Neural Network for Low Cost Distorted Image Classification

Deep convolutional neural networks have proven to be well suited for ima...
research
08/22/2023

Using Early Exits for Fast Inference in Automatic Modulation Classification

Automatic modulation classification (AMC) plays a critical role in wirel...
research
10/03/2022

Decompiling x86 Deep Neural Network Executables

Due to their widespread use on heterogeneous hardware devices, deep lear...
research
02/27/2017

Fast and Accurate Inference with Adaptive Ensemble Prediction in Image Classification with Deep Neural Networks

Ensembling multiple predictions is a widely used technique to improve th...
research
06/07/2018

Path-Level Network Transformation for Efficient Architecture Search

We introduce a new function-preserving transformation for efficient neur...
research
09/29/2022

Data Flex: On-Platform Organisations

The natural alignment between business and architecture within big techs...

Please sign up or login with your details

Forgot password? Click here to reset