Review: Ordinary Differential Equations For Deep Learning

11/01/2019
by   Xinshi Chen, et al.
0

To better understand and improve the behavior of neural networks, a recent line of works bridged the connection between ordinary differential equations (ODEs) and deep neural networks (DNNs). The connections are made in two folds: (1) View DNN as ODE discretization; (2) View the training of DNN as solving an optimal control problem. The former connection motivates people either to design neural architectures based on ODE discretization schemes or to replace DNN by a continuous model characterized by ODEs. Several works demonstrated distinct advantages of using a continuous model instead of traditional DNN in some specific applications. The latter connection is inspiring. Based on Pontryagin's maximum principle, which is popular in the optimal control literature, some developed new optimization methods for training neural networks and some developed algorithms to train the infinite-deep continuous model with low memory-cost. This paper is organized as follows: In Section 2, the relation between neural architecture and ODE discretization is introduced. Some architectures are not motivated by ODE, but they are later found to be associated with some specific discretization schemes. Some architectures are designed based on ODE discretization and expected to achieve some special properties. Section 3 formulates the optimization problem where a traditional neural network is replaced by a continuous model (ODE). The formulated optimization problem is an optimal control problem. Therefore, two different types of controls will also be discussed in this section. In Section 4, we will discuss how we can utilize the optimization methods that are popular in optimal control literature to help the training of machine learning problems. Finally, two applications of using a continuous model will be shown in Section 5 and 6 to demonstrate some of its advantages over traditional neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/19/2019

Numerical Optimal Control of HIV Transmission in Octave/MATLAB

We provide easy and readable GNU Octave/MATLAB code for the simulation o...
research
07/05/2020

Depth-Adaptive Neural Networks from the Optimal Control viewpoint

In recent years, deep learning has been connected with optimal control a...
research
03/22/2023

Leveraging Multi-time Hamilton-Jacobi PDEs for Certain Scientific Machine Learning Problems

Hamilton-Jacobi partial differential equations (HJ PDEs) have deep conne...
research
05/08/2021

Dynamic Game Theoretic Neural Optimizer

The connection between training deep neural networks (DNNs) and optimal ...
research
11/17/2015

Neurocontrol methods review

Methods of applying neural networks to control plants are considered. Me...
research
01/08/2021

On the Turnpike to Design of Deep Neural Nets: Explicit Depth Bounds

It is well-known that the training of Deep Neural Networks (DNN) can be ...
research
07/26/2020

Train Like a (Var)Pro: Efficient Training of Neural Networks with Variable Projection

Deep neural networks (DNNs) have achieved state-of-the-art performance a...

Please sign up or login with your details

Forgot password? Click here to reset