Neural signature kernels as infinite-width-depth-limits of controlled ResNets

03/30/2023
by   Nicola Muca Cirone, et al.
0

Motivated by the paradigm of reservoir computing, we consider randomly initialized controlled ResNets defined as Euler-discretizations of neural controlled differential equations (Neural CDEs). We show that in the infinite-width-then-depth limit and under proper scaling, these architectures converge weakly to Gaussian processes indexed on some spaces of continuous paths and with kernels satisfying certain partial differential equations (PDEs) varying according to the choice of activation function. In the special case where the activation is the identity, we show that the equation reduces to a linear PDE and the limiting kernel agrees with the signature kernel of Salvi et al. (2021). In this setting, we also show that the width-depth limits commute. We name this new family of limiting kernels neural signature kernels. Finally, we show that in the infinite-depth regime, finite-width controlled ResNets converge in distribution to Neural CDEs with random vector fields which, depending on whether the weights are shared across layers, are either time-independent and Gaussian or behave like a matrix-valued Brownian motion.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/03/2022

On the infinite-depth limit of finite-width neural networks

In this paper, we study the infinite-depth limit of finite-width residua...
research
06/06/2022

The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization

The logit outputs of a feedforward neural network at initialization are ...
research
08/10/2020

Do ideas have shape? Plato's theory of forms as the continuous limit of artificial neural networks

We show that ResNets converge, in the infinite depth limit, to a general...
research
12/01/2021

Mixed neural network Gaussian processes

This paper makes two contributions. Firstly, it introduces mixed composi...
research
10/10/2022

Meta-Principled Family of Hyperparameter Scaling Strategies

In this note, we first derive a one-parameter family of hyperparameter s...
research
09/26/2022

A connection between probability, physics and neural networks

We illustrate an approach that can be exploited for constructing neural ...
research
06/30/2023

The Shaped Transformer: Attention Models in the Infinite Depth-and-Width Limit

In deep learning theory, the covariance matrix of the representations se...

Please sign up or login with your details

Forgot password? Click here to reset