Unsupervised Pretraining Encourages Moderate-Sparseness

12/20/2013
by   Jun Li, et al.
0

It is well known that direct training of deep neural networks will generally lead to poor results. A major progress in recent years is the invention of various pretraining methods to initialize network parameters and it was shown that such methods lead to good prediction performance. However, the reason for the success of pretraining has not been fully understood, although it was argued that regularization and better optimization play certain roles. This paper provides another explanation for the effectiveness of pretraining, where we show pretraining leads to a sparseness of hidden unit activation in the resulting neural networks. The main reason is that the pretraining models can be interpreted as an adaptive sparse coding. Compared to deep neural network with sigmoid function, our experimental results on MNIST and Birdsong further support this sparseness observation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2016

Multi-pretrained Deep Neural Network

Pretraining is widely used in deep neutral network and one of the most f...
research
10/28/2022

LOFT: Finding Lottery Tickets through Filter-wise Training

Recent work on the Lottery Ticket Hypothesis (LTH) shows that there exis...
research
02/27/2023

Linear pretraining in recurrent mixture density networks

We present a method for pretraining a recurrent mixture density network ...
research
06/08/2015

Adaptive Normalized Risk-Averting Training For Deep Neural Networks

This paper proposes a set of new error criteria and learning approaches,...
research
12/01/2021

Revisiting the Transferability of Supervised Pretraining: an MLP Perspective

The pretrain-finetune paradigm is a classical pipeline in visual learnin...
research
03/07/2023

Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?

Pretraining a neural network on a large dataset is becoming a cornerston...
research
06/07/2017

Are Saddles Good Enough for Deep Learning?

Recent years have seen a growing interest in understanding deep neural n...

Please sign up or login with your details

Forgot password? Click here to reset