Neural Network Layer Matrix Decomposition reveals Latent Manifold Encoding and Memory Capacity

09/12/2023
by   Ng Shyh-Chang, et al.
0

We prove the converse of the universal approximation theorem, i.e. a neural network (NN) encoding theorem which shows that for every stably converged NN of continuous activation functions, its weight matrix actually encodes a continuous function that approximates its training dataset to within a finite margin of error over a bounded domain. We further show that using the Eckart-Young theorem for truncated singular value decomposition of the weight matrix for every NN layer, we can illuminate the nature of the latent space manifold of the training dataset encoded and represented by every NN layer, and the geometric nature of the mathematical operations performed by each NN layer. Our results have implications for understanding how NNs break the curse of dimensionality by harnessing memory capacity for expressivity, and that the two are complementary. This Layer Matrix Decomposition (LMD) further suggests a close relationship between eigen-decomposition of NN layers and the latest advances in conceptualizations of Hopfield networks and Transformer NN models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/30/2021

Two Instances of Interpretable Neural Network for Universal Approximations

This paper proposes two bottom-up interpretable neural network (NN) cons...
research
06/25/2021

Tensor-based framework for training flexible neural networks

Activation functions (AFs) are an important part of the design of neural...
research
09/22/2020

Tensor Programs III: Neural Matrix Laws

In a neural network (NN), weight matrices linearly transform inputs into...
research
11/25/2022

LU decomposition and Toeplitz decomposition of a neural network

It is well-known that any matrix A has an LU decomposition. Less well-kn...
research
01/10/2023

Optimal Power Flow Based on Physical-Model-Integrated Neural Network with Worth-Learning Data Generation

Fast and reliable solvers for optimal power flow (OPF) problems are attr...
research
04/10/2020

Entropy-Based Modeling for Estimating Soft Errors Impact on Binarized Neural Network Inference

Over past years, the easy accessibility to the large scale datasets has ...
research
06/11/2023

Precise and Generalized Robustness Certification for Neural Networks

The objective of neural network (NN) robustness certification is to dete...

Please sign up or login with your details

Forgot password? Click here to reset