Universal Solutions of Feedforward ReLU Networks for Interpolations

08/16/2022
by   Changcun Huang, et al.
0

This paper provides a theoretical framework on the solution of feedforward ReLU networks for interpolations, in terms of what is called an interpolation matrix, which is the summary, extension and generalization of our three preceding works, with the expectation that the solution of engineering could be included in this framework and finally understood. To three-layer networks, we classify different kinds of solutions and model them in a normalized form; the solution finding is investigated by three dimensions, including data, networks and the training; the mechanism of overparameterization solutions is interpreted. To deep-layer networks, we present a general result called sparse-matrix principle, which could describe some basic behavior of deep layers and explain the phenomenon of the sparse-activation mode that appears in engineering applications associated with brain science; an advantage of deep layers compared to shallower ones is manifested in this principle. As applications, a general solution of deep neural networks for classification is constructed by that principle; and we also use the principle to study the data-disentangling property of encoders. Analogous to the three-layer case, the solution of deep layers is also explored through several dimensions. The mechanism of multi-output neural networks is explained from the perspective of interpolation matrices.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2020

Convex Geometry and Duality of Over-parameterized Neural Networks

We develop a convex analytic framework for ReLU neural networks which el...
research
12/12/2022

On an Interpretation of ResNets via Solution Constructions

This paper first constructs a typical solution of ResNets for multi-cate...
research
08/15/2022

On a Mechanism Framework of Autoencoders

This paper proposes a theoretical framework on the mechanism of autoenco...
research
04/24/2022

Piecewise-Linear Activations or Analytic Activation Functions: Which Produce More Expressive Neural Networks?

Many currently available universal approximation theorems affirm that de...
research
02/22/2020

Convex Duality of Deep Neural Networks

We study regularized deep neural networks and introduce an analytic fram...
research
03/18/2019

On-line learning dynamics of ReLU neural networks using statistical physics techniques

We introduce exact macroscopic on-line learning dynamics of two-layer ne...
research
06/12/2019

Evaluation of Dataflow through layers of Deep Neural Networks in Classification and Regression Problems

This paper introduces two straightforward, effective indices to evaluate...

Please sign up or login with your details

Forgot password? Click here to reset