A Shooting Formulation of Deep Learning

Continuous-depth neural networks can be viewed as deep limits of discrete neural networks whose dynamics resemble a discretization of an ordinary differential equation (ODE). Although important steps have been taken to realize the advantages of such continuous formulations, most current techniques are not truly continuous-depth as they assume identical layers. Indeed, existing works throw into relief the myriad difficulties presented by an infinite-dimensional parameter space in learning a continuous-depth neural ODE. To this end, we introduce a shooting formulation which shifts the perspective from parameterizing a network layer-by-layer to parameterizing over optimal networks described only by a set of initial conditions. For scalability, we propose a novel particle-ensemble parametrization which fully specifies the optimal weight trajectory of the continuous-depth neural network. Our experiments show that our particle-ensemble shooting formulation can achieve competitive performance, especially on long-range forecasting tasks. Finally, though the current work is inspired by continuous-depth neural networks, the particle-ensemble shooting formulation also applies to discrete-time networks and may lead to a new fertile area of research in deep learning parametrization.

READ FULL TEXT

page 9

page 10

page 27

page 28

research
07/05/2020

Depth-Adaptive Neural Networks from the Optimal Control viewpoint

In recent years, deep learning has been connected with optimal control a...
research
09/20/2020

TorchDyn: A Neural Differential Equations Library

Continuous-depth learning has recently emerged as a novel perspective on...
research
07/19/2020

Hypersolvers: Toward Fast Continuous-Depth Models

The infinite-depth paradigm pioneered by Neural ODEs has launched a rena...
research
08/18/2019

Neural Dynamics on Complex Networks

We introduce a deep learning model to learn continuous-time dynamics on ...
research
01/06/2022

Nonlocal Kernel Network (NKN): a Stable and Resolution-Independent Deep Neural Network

Neural operators have recently become popular tools for designing soluti...
research
07/05/2017

Machine Learning, Deepest Learning: Statistical Data Assimilation Problems

We formulate a strong equivalence between machine learning, artificial i...
research
11/21/2019

Discrete and Continuous Deep Residual Learning Over Graphs

In this paper we propose the use of continuous residual modules for grap...

Please sign up or login with your details

Forgot password? Click here to reset