Training Sparse Neural Networks

11/21/2016
by   Suraj Srinivas, et al.
0

Deep neural networks with lots of parameters are typically used for large-scale computer vision tasks such as image classification. This is a result of using dense matrix multiplications and convolutions. However, sparse computations are known to be much more efficient. In this work, we train and build neural networks which implicitly use sparse computations. We introduce additional gate variables to perform parameter selection and show that this is equivalent to using a spike-and-slab prior. We experimentally validate our method on both small and large networks and achieve state-of-the-art compression results for sparse neural network models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2016

Improving the Robustness of Deep Neural Networks via Stability Training

In this paper we address the issue of output instability of deep neural ...
research
04/15/2018

SparseNet: A Sparse DenseNet for Image Classification

Deep neural networks have made remarkable progresses on various computer...
research
01/22/2021

Selfish Sparse RNN Training

Sparse neural networks have been widely applied to reduce the necessary ...
research
12/09/2022

Sparse Upcycling: Training Mixture-of-Experts from Dense Checkpoints

Training large, deep neural networks to convergence can be prohibitively...
research
08/14/2021

A Sparse Coding Interpretation of Neural Networks and Theoretical Implications

Neural networks, specifically deep convolutional neural networks, have a...
research
06/21/2022

Winning the Lottery Ahead of Time: Efficient Early Network Pruning

Pruning, the task of sparsifying deep neural networks, received increasi...
research
07/10/2023

SparseVSR: Lightweight and Noise Robust Visual Speech Recognition

Recent advances in deep neural networks have achieved unprecedented succ...

Please sign up or login with your details

Forgot password? Click here to reset