On the rate of convergence of a neural network regression estimate learned by gradient descent

12/09/2019
by   Alina Braun, et al.
0

Nonparametric regression with random design is considered. Estimates are defined by minimzing a penalized empirical L_2 risk over a suitably chosen class of neural networks with one hidden layer via gradient descent. Here, the gradient descent procedure is repeated several times with randomly chosen starting values for the weights, and from the list of constructed estimates the one with the minimal empirical L_2 risk is chosen. Under the assumption that the number of randomly chosen starting values and the number of steps for gradient descent are sufficiently large it is shown that the resulting estimate achieves (up to a logarithmic factor) the optimal rate of convergence in a projection pursuit model. The final sample size performance of the estimates is illustrated by using simulated data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/09/2019

Analysis of the rate of convergence of neural network regression estimates which are easy to implement

Recent results in nonparametric regression show that for deep learning, ...
research
07/20/2021

The Smoking Gun: Statistical Theory Improves Neural Network Estimates

In this paper we analyze the L_2 error of neural network regression esti...
research
08/30/2022

On the universal consistency of an over-parametrized deep neural network estimate learned by gradient descent

Estimation of a multivariate regression function from independent and id...
research
10/04/2022

Analysis of the rate of convergence of an over-parametrized deep neural network estimate learned by gradient descent

Estimation of a regression function from independent and identically dis...
research
11/07/2019

How implicit regularization of Neural Networks affects the learned function – Part I

Today, various forms of neural networks are trained to perform approxima...
research
05/07/2018

Polynomial Convergence of Gradient Descent for Training One-Hidden-Layer Neural Networks

We analyze Gradient Descent applied to learning a bounded target functio...
research
07/19/2019

Surfing: Iterative optimization over incrementally trained deep networks

We investigate a sequential optimization procedure to minimize the empir...

Please sign up or login with your details

Forgot password? Click here to reset