Mehler's Formula, Branching Process, and Compositional Kernels of Deep Neural Networks

04/09/2020
by   Tengyuan Liang, et al.
0

In this paper, we utilize a connection between compositional kernels and branching processes via Mehler's formula to study deep neural networks. This new probabilistic insight provides us a novel perspective on the mathematical role of activation functions in compositional neural networks. We study the unscaled and rescaled limits of the compositional kernels and explore the different phases of the limiting behavior, as the compositional depth increases. We investigate the memorization capacity of the compositional kernels and neural networks by characterizing the interplay among compositional depth, sample size, dimensionality, and non-linearity of the activation. Explicit formulas on the eigenvalues of the compositional kernel are provided, which quantify the complexity of the corresponding reproducing kernel Hilbert space. On the algorithmic front, we propose a new random features algorithm, which compresses the compositional layers by devising a new activation function.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/01/2021

Mixed neural network Gaussian processes

This paper makes two contributions. Firstly, it introduces mixed composi...
research
03/04/2020

Neural Kernels Without Tangents

We investigate the connections between neural networks and simple buildi...
research
05/02/2022

Reproducing Kernels and New Approaches in Compositional Data Analysis

Compositional data, such as human gut microbiomes, consist of non-negati...
research
02/17/2020

Investigating the Compositional Structure Of Deep Neural Networks

The current understanding of deep neural networks can only partially exp...
research
02/18/2016

Toward Deeper Understanding of Neural Networks: The Power of Initialization and a Dual View on Expressivity

We develop a general duality between neural networks and compositional k...
research
05/23/2023

Evolution: A Unified Formula for Feature Operators from a High-level Perspective

Traditionally, different types of feature operators (e.g., convolution, ...
research
10/11/2019

The Expressivity and Training of Deep Neural Networks: toward the Edge of Chaos?

Expressivity is one of the most significant issues in assessing neural n...

Please sign up or login with your details

Forgot password? Click here to reset