Theoretical Properties for Neural Networks with Weight Matrices of Low Displacement Rank

03/01/2017
by   Liang Zhao, et al.
0

Recently low displacement rank (LDR) matrices, or so-called structured matrices, have been proposed to compress large-scale neural networks. Empirical results have shown that neural networks with weight matrices of LDR matrices, referred as LDR neural networks, can achieve significant reduction in space and computational complexity while retaining high accuracy. We formally study LDR matrices in deep learning. First, we prove the universal approximation property of LDR neural networks with a mild condition on the displacement operators. We then show that the error bounds of LDR neural networks are as efficient as general neural networks with both single-layer and multiple-layer structure. Finally, we propose back-propagation based training algorithm for general LDR neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/04/2018

Learning Compressed Transforms with Low Displacement Rank

The low displacement rank (LDR) framework for structured matrices repres...
research
07/21/2018

On the Analysis of Trajectories of Gradient Descent in the Optimization of Deep Neural Networks

Theoretical analysis of the error landscape of deep neural networks has ...
research
09/13/2019

Electro-optical Neural Networks based on Time-stretch Method

In this paper, a novel architecture of electro-optical neural networks b...
research
10/27/2003

Feedforward Neural Networks with Diffused Nonlinear Weight Functions

In this paper, feedforward neural networks are presented that have nonli...
research
10/28/2017

Trainable back-propagated functional transfer matrices

Connections between nodes of fully connected neural networks are usually...
research
11/01/2021

Investigating the locality of neural network training dynamics

A fundamental quest in the theory of deep-learning is to understand the ...
research
01/29/2019

On the Expressive Power of Deep Fully Circulant Neural Networks

In this paper, we study deep fully circulant neural networks, that is de...

Please sign up or login with your details

Forgot password? Click here to reset