Machine Learning, Deepest Learning: Statistical Data Assimilation Problems

07/05/2017
by   Henry Abarbanel, et al.
0

We formulate a strong equivalence between machine learning, artificial intelligence methods and the formulation of statistical data assimilation as used widely in physical and biological sciences. The correspondence is that layer number in the artificial network setting is the analog of time in the data assimilation setting. Within the discussion of this equivalence we show that adding more layers (making the network deeper) is analogous to adding temporal resolution in a data assimilation framework. How one can find a candidate for the global minimum of the cost functions in the machine learning context using a method from data assimilation is discussed. Calculations on simple models from each side of the equivalence are reported. Also discussed is a framework in which the time or layer label is taken to be continuous, providing a differential equation, the Euler-Lagrange equation, which shows that the problem being solved is a two point boundary value problem familiar in the discussion of variational methods. The use of continuous layers is denoted "deepest learning". These problems respect a symplectic symmetry in continuous time/layer phase space. Both Lagrangian versions and Hamiltonian versions of these problems are presented. Their well-studied implementation in a discrete time/layer, while respected the symplectic structure, is addressed. The Hamiltonian version provides a direct rationale for back propagation as a solution method for the canonical momentum.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2017

Machine Learning as Statistical Data Assimilation

We identify a strong equivalence between neural network based machine le...
research
03/21/2022

Continuous-Stage Runge-Kutta approximation to Differential Problems

In recent years, the efficient numerical solution of Hamiltonian problem...
research
06/17/2022

Learning the parameters of a differential equation from its trajectory via the adjoint equation

The paper contributes to strengthening the relation between machine lear...
research
10/27/2020

Hamilton-Jacobi Deep Q-Learning for Deterministic Continuous-Time Systems with Lipschitz Continuous Controls

In this paper, we propose Q-learning algorithms for continuous-time dete...
research
06/18/2020

A Shooting Formulation of Deep Learning

Continuous-depth neural networks can be viewed as deep limits of discret...
research
02/10/2018

On Symplectic Optimization

Accelerated gradient methods have had significant impact in machine lear...

Please sign up or login with your details

Forgot password? Click here to reset