PCA-Initialized Deep Neural Networks Applied To Document Image Analysis

02/01/2017
by   Mathias Seuret, et al.
0

In this paper, we present a novel approach for initializing deep neural networks, i.e., by turning PCA into neural layers. Usually, the initialization of the weights of a deep neural network is done in one of the three following ways: 1) with random values, 2) layer-wise, usually as Deep Belief Network or as auto-encoder, and 3) re-use of layers from another network (transfer learning). Therefore, typically, many training epochs are needed before meaningful weights are learned, or a rather similar dataset is required for seeding a fine-tuning of transfer learning. In this paper, we describe how to turn a PCA into an auto-encoder, by generating an encoder layer of the PCA parameters and furthermore adding a decoding layer. We analyze the initialization technique on real documents. First, we show that a PCA-based initialization is quick and leads to a very stable initialization. Furthermore, for the task of layout analysis we investigate the effectiveness of PCA-based initialization and show that it outperforms state-of-the-art random weight initialization methods.

READ FULL TEXT

page 3

page 4

page 5

research
10/19/2017

Historical Document Image Segmentation with LDA-Initialized Deep Neural Networks

In this paper, we present a novel approach to perform deep neural networ...
research
09/18/2019

A Study on Binary Neural Networks Initialization

Initialization plays a crucial role in training neural models. Binary Ne...
research
07/18/2022

Explainable Deep Belief Network based Auto encoder using novel Extended Garson Algorithm

The most difficult task in machine learning is to interpret trained shal...
research
05/02/2021

Data-driven Weight Initialization with Sylvester Solvers

In this work, we propose a data-driven scheme to initialize the paramete...
research
08/08/2020

Using UNet and PSPNet to explore the reusability principle of CNN parameters

How to reduce the requirement on training dataset size is a hot topic in...
research
02/02/2018

Intriguing Properties of Randomly Weighted Networks: Generalizing While Learning Next to Nothing

Training deep neural networks results in strong learned representations ...
research
04/16/2020

Hcore-Init: Neural Network Initialization based on Graph Degeneracy

Neural networks are the pinnacle of Artificial Intelligence, as in recen...

Please sign up or login with your details

Forgot password? Click here to reset