The Outer Product Structure of Neural Network Derivatives

10/09/2018
by   Craig Bakker, et al.
0

In this paper, we show that feedforward and recurrent neural networks exhibit an outer product derivative structure but that convolutional neural networks do not. This structure makes it possible to use higher-order information without needing approximations or infeasibly large amounts of memory, and it may also provide insights into the geometry of neural network optima. The ability to easily access these derivatives also suggests a new, geometric approach to regularization. We then discuss how this structure could be used to improve training methods, increase network robustness and generalizability, and inform network compression methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/15/2016

A Geometric Framework for Convolutional Neural Networks

In this paper, a geometric framework for neural networks is proposed. Th...
research
07/17/2018

Expressive power of outer product manifolds on feed-forward neural networks

Hierarchical neural networks are exponentially more efficient than their...
research
06/03/2019

Factor Graph Neural Network

Most of the successful deep neural network architectures are structured,...
research
08/13/2018

Rank-1 Convolutional Neural Network

In this paper, we propose a convolutional neural network(CNN) with 3-D r...
research
08/18/2020

An a posteriori error estimate of the outer normal derivative using dual weights

We derive a residual based a-posteriori error estimate for the outer nor...
research
10/18/2021

Speeding-Up Back-Propagation in DNN: Approximate Outer Product with Memory

In this paper, an algorithm for approximate evaluation of back-propagati...

Please sign up or login with your details

Forgot password? Click here to reset