Analysis of the rate of convergence of an over-parametrized deep neural network estimate learned by gradient descent

10/04/2022
by   Michael Kohler, et al.
0

Estimation of a regression function from independent and identically distributed random variables is considered. The L_2 error with integration with respect to the design measure is used as an error criterion. Over-parametrized deep neural network estimates are defined where all the weights are learned by the gradient descent. It is shown that the expected L_2 error of these estimates converges to zero with the rate close to n^-1/(1+d) in case that the regression function is Hölder smooth with Hölder exponent p ∈ [1/2,1]. In case of an interaction model where the regression function is assumed to be a sum of Hölder smooth functions where each of the functions depends only on d^* many of d components of the design variable, it is shown that these estimates achieve the corresponding d^*-dimensional rate of convergence.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2021

Estimation of a regression function on a manifold by fully connected deep neural networks

Estimation of a regression function from independent and identically dis...
research
08/30/2022

On the universal consistency of an over-parametrized deep neural network estimate learned by gradient descent

Estimation of a multivariate regression function from independent and id...
research
07/20/2021

The Smoking Gun: Statistical Theory Improves Neural Network Estimates

In this paper we analyze the L_2 error of neural network regression esti...
research
12/09/2019

Analysis of the rate of convergence of neural network regression estimates which are easy to implement

Recent results in nonparametric regression show that for deep learning, ...
research
04/17/2023

Pointwise convergence theorem of generalized mini-batch gradient descent in deep neural network

The theoretical structure of deep neural network (DNN) has been clarifie...
research
12/09/2019

On the rate of convergence of a neural network regression estimate learned by gradient descent

Nonparametric regression with random design is considered. Estimates are...
research
04/10/2019

Analysis of the Gradient Descent Algorithm for a Deep Neural Network Model with Skip-connections

The behavior of the gradient descent (GD) algorithm is analyzed for a de...

Please sign up or login with your details

Forgot password? Click here to reset