Explore the Knowledge contained in Network Weights to Obtain Sparse Neural Networks

03/26/2021
by   Mengqiao Han, et al.
0

Sparse neural networks are important for achieving better generalization and enhancing computation efficiency. This paper proposes a novel learning approach to obtain sparse fully connected layers in neural networks (NNs) automatically. We design a switcher neural network (SNN) to optimize the structure of the task neural network (TNN). The SNN takes the weights of the TNN as the inputs and its outputs are used to switch the connections of TNN. In this way, the knowledge contained in the weights of TNN is explored to determine the importance of each connection and the structure of TNN consequently. The SNN and TNN are learned alternately with stochastic gradient descent (SGD) optimization, targeting at a common objective. After learning, we achieve the optimal structure and the optimal parameters of the TNN simultaneously. In order to evaluate the proposed approach, we conduct image classification experiments on various network structures and datasets. The network structures include LeNet, ResNet18, ResNet34, VggNet16 and MobileNet. The datasets include MNIST, CIFAR10 and CIFAR100. The experimental results show that our approach can stably lead to sparse and well-performing fully connected layers in NNs.

READ FULL TEXT
research
10/11/2021

Mining the Weights Knowledge for Optimizing Neural Network Structures

Knowledge embedded in the weights of the artificial neural network can b...
research
07/15/2019

Controlling Model Complexity in Probabilistic Model-Based Dynamic Optimization of Neural Network Structures

A method of simultaneously optimizing both the structure of neural netwo...
research
11/06/2017

Characterizing Sparse Connectivity Patterns in Neural Networks

We propose a novel way of reducing the number of parameters in the stora...
research
03/01/2021

Deep Learning with a Classifier System: Initial Results

This article presents the first results from using a learning classifier...
research
06/16/2021

Structured DropConnect for Uncertainty Inference in Image Classification

With the complexity of the network structure, uncertainty inference has ...
research
05/09/2020

GPU Acceleration of Sparse Neural Networks

In this paper, we use graphics processing units(GPU) to accelerate spars...
research
05/27/2021

Learning Structures for Deep Neural Networks

In this paper, we focus on the unsupervised setting for structure learni...

Please sign up or login with your details

Forgot password? Click here to reset