Random Projection in Deep Neural Networks

12/22/2018
by   Piotr Iwo Wójcik, et al.
0

This work investigates the ways in which deep learning methods can benefit from random projection (RP), a classic linear dimensionality reduction method. We focus on two areas where, as we have found, employing RP techniques can improve deep models: training neural networks on high-dimensional data and initialization of network parameters. Training deep neural networks (DNNs) on sparse, high-dimensional data with no exploitable structure implies a network architecture with an input layer that has a huge number of weights, which often makes training infeasible. We show that this problem can be solved by prepending the network with an input layer whose weights are initialized with an RP matrix. We propose several modifications to the network architecture and training regime that makes it possible to efficiently train DNNs with learnable RP layer on data with as many as tens of millions of input features and training examples. In comparison to the state-of-the-art methods, neural networks with RP layer achieve competitive performance or improve the results on several extremely high-dimensional real-world datasets. The second area where the application of RP techniques can be beneficial for training deep models is weight initialization. Setting the initial weights in DNNs to elements of various RP matrices enabled us to train residual deep networks to higher levels of performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2017

Weight Initialization of Deep Neural Networks(DNNs) using Data Statistics

Deep neural networks (DNNs) form the backbone of almost every state-of-t...
research
12/08/2022

SpaceEditing: Integrating Human Knowledge into Deep Neural Networks via Interactive Latent Space Editing

We propose an interactive editing method that allows humans to help deep...
research
11/03/2020

Parameter Efficient Deep Neural Networks with Bilinear Projections

Recent research on deep neural networks (DNNs) has primarily focused on ...
research
07/20/2016

On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks

Next generation deep neural networks for classification hosted on embedd...
research
05/11/2020

Ensembled sparse-input hierarchical networks for high-dimensional datasets

Neural networks have seen limited use in prediction for high-dimensional...
research
12/19/2019

Multilevel Initialization for Layer-Parallel Deep Neural Network Training

This paper investigates multilevel initialization strategies for trainin...

Please sign up or login with your details

Forgot password? Click here to reset