Differential equations as models of deep neural networks

09/09/2019
by   Julius Ruseckas, et al.
0

In this work we systematically analyze general properties of differential equations used as machine learning models. We demonstrate that the gradient of the loss function with respect to to the hidden state can be considered as a generalized momentum conjugate to the hidden state, allowing application of the tools of classical mechanics. In addition, we show that not only residual networks, but also feedforward neural networks with small nonlinearities and the weights matrices deviating only slightly from identity matrices can be related to the differential equations. We propose a differential equation describing such networks and investigate its properties.

READ FULL TEXT
research
02/06/2019

DiffEqFlux.jl - A Julia Library for Neural Differential Equations

DiffEqFlux.jl is a library for fusing neural networks and differential e...
research
05/23/2022

Learning differential equations from data

Differential equations are used to model problems that originate in disc...
research
09/22/2022

Vanilla feedforward neural networks as a discretization of dynamic systems

Deep learning has made significant applications in the field of data sci...
research
09/20/2021

Machine-learning hidden symmetries

We present an automated method for finding hidden symmetries, defined as...
research
11/01/2018

How the fundamental concepts of mathematics and physics explain deep learning

Starting from the Fermat's principle of least action, which governs clas...
research
05/27/2019

Neural Stochastic Differential Equations

Deep neural networks whose parameters are distributed according to typic...
research
08/03/2021

Numerical Solution of Stiff Ordinary Differential Equations with Random Projection Neural Networks

We propose a numerical scheme based on Random Projection Neural Networks...

Please sign up or login with your details

Forgot password? Click here to reset