Stochastic Gradient Descent in Continuous Time

11/17/2016
by   Justin Sirignano, et al.
0

Stochastic gradient descent in continuous time (SGDCT) provides a computationally efficient method for the statistical learning of continuous-time models, which are widely used in science, engineering, and finance. The SGDCT algorithm follows a (noisy) descent direction along a continuous stream of data. SGDCT performs an online parameter update in continuous time, with the parameter updates θ_t satisfying a stochastic differential equation. We prove that _t →∞∇g̅(θ_t) = 0 where g̅ is a natural objective function for the estimation of the continuous-time dynamics. The convergence proof leverages ergodicity by using an appropriate Poisson equation to help describe the evolution of the parameters for large times. SGDCT can also be used to solve continuous-time optimization problems, such as American options. For certain continuous-time problems, SGDCT has some promising advantages compared to a traditional stochastic gradient descent algorithm. As an example application, SGDCT is combined with a deep neural network to price high-dimensional American options (up to 100 dimensions).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/11/2017

Stochastic Gradient Descent in Continuous Time: A Central Limit Theorem

Stochastic gradient descent in continuous time (SGDCT) provides a comput...
research
08/19/2020

SODEN: A Scalable Continuous-Time Survival Model through Ordinary Differential Equation Networks

In this paper, we propose a flexible model for survival analysis using n...
research
07/11/2023

Implicit regularisation in stochastic gradient descent: from single-objective to two-player games

Recent years have seen many insights on deep learning optimisation being...
research
06/04/2021

Stochastic gradient descent with noise of machine learning type. Part II: Continuous time analysis

The representation of functions by artificial neural networks depends on...
research
10/04/2018

Convergence of the ADAM algorithm from a Dynamical System Viewpoint

Adam is a popular variant of the stochastic gradient descent for finding...
research
10/28/2022

Toward Equation of Motion for Deep Neural Networks: Continuous-time Gradient Descent and Discretization Error Analysis

We derive and solve an “Equation of Motion” (EoM) for deep neural networ...
research
02/10/2022

On One-Bit Quantization

We consider the one-bit quantizer that minimizes the mean squared error ...

Please sign up or login with your details

Forgot password? Click here to reset