Infinitely wide limits for deep Stable neural networks: sub-linear, linear and super-linear activation functions

04/08/2023
by   Alberto Bordino, et al.
0

There is a growing literature on the study of large-width properties of deep Gaussian neural networks (NNs), i.e. deep NNs with Gaussian-distributed parameters or weights, and Gaussian stochastic processes. Motivated by some empirical and theoretical studies showing the potential of replacing Gaussian distributions with Stable distributions, namely distributions with heavy tails, in this paper we investigate large-width properties of deep Stable NNs, i.e. deep NNs with Stable-distributed parameters. For sub-linear activation functions, a recent work has characterized the infinitely wide limit of a suitable rescaled deep Stable NN in terms of a Stable stochastic process, both under the assumption of a “joint growth" and under the assumption of a “sequential growth" of the width over the NN's layers. Here, assuming a “sequential growth" of the width, we extend such a characterization to a general class of activation functions, which includes sub-linear, asymptotically linear and super-linear functions. As a novelty with respect to previous works, our results rely on the use of a generalized central limit theorem for heavy tails distributions, which allows for an interesting unified treatment of infinitely wide limits for deep Stable NNs. Our study shows that the scaling of Stable NNs and the stability of their infinitely wide limits may depend on the choice of the activation function, bringing out a critical difference with respect to the Gaussian setting.

READ FULL TEXT

page 19

page 20

research
06/16/2022

Neural tangent kernel analysis of shallow α-Stable ReLU neural networks

There is a recent literature on large-width properties of Gaussian neura...
research
08/02/2021

Deep Stable neural networks: large-width asymptotics and convergence rates

In modern deep learning, there is a recent and growing literature on the...
research
03/01/2020

Stable behaviour of infinitely wide deep neural networks

We consider fully connected feed-forward deep neural networks (NNs) wher...
research
09/26/2022

A connection between probability, physics and neural networks

We illustrate an approach that can be exploited for constructing neural ...
research
06/18/2021

α-Stable convergence of heavy-tailed infinitely-wide neural networks

We consider infinitely-wide multi-layer perceptrons (MLPs) which are lim...
research
02/07/2021

Infinite-channel deep stable convolutional neural networks

The interplay between infinite-width neural networks (NNs) and classes o...
research
06/10/2020

Scalable Partial Explainability in Neural Networks via Flexible Activation Functions

Achieving transparency in black-box deep learning algorithms is still an...

Please sign up or login with your details

Forgot password? Click here to reset