Kinetic Theory for Residual Neural Networks

01/07/2020
by   M. Herty, et al.
0

Deep residual neural networks (ResNet) are performing very well for many data science applications. We use kinetic theory to improve understanding and existing methods. A microscopic simplified residual neural network (SimResNet) model is studied as the limit of infinitely many inputs. This leads to kinetic formulations of the SimResNet and we analyze those with respect to sensitivities and steady states. Aggregation phenomena in the case of a linear activation function are also studied. In addition the analysis is validated by numerics. In particular, results on a clustering and regression problem are presented.

READ FULL TEXT

page 14

page 18

research
07/07/2020

Towards an Understanding of Residual Networks Using Neural Tangent Hierarchy (NTH)

Gradient descent yields zero training loss in polynomial time for deep n...
research
01/08/2021

Residual networks classify inputs based on their neural transient dynamics

In this study, we analyze the input-output behavior of residual networks...
research
05/11/2020

Simplified ResNet approach for data driven prediction of microstructure-fatigue relationship

The heterogeneous microstructure in metallic components results in local...
research
10/15/2020

Error estimates of residual minimization using neural networks for linear PDEs

We propose an abstract framework for analyzing the convergence of least-...
research
09/21/2020

Kernel-Based Smoothness Analysis of Residual Networks

A major factor in the success of deep neural networks is the use of soph...
research
05/27/2019

Identity Connections in Residual Nets Improve Noise Stability

Residual Neural Networks (ResNets) achieve state-of-the-art performance ...
research
02/15/2021

A Koopman Approach to Understanding Sequence Neural Models

We introduce a new approach to understanding trained sequence neural mod...

Please sign up or login with your details

Forgot password? Click here to reset