Nonparametric regression using over-parameterized shallow ReLU neural networks

06/14/2023
by   Yunfei Yang, et al.
0

It is shown that over-parameterized neural networks can achieve minimax optimal rates of convergence (up to logarithmic factors) for learning functions from certain smooth function classes, if the weights are suitably constrained or regularized. Specifically, we consider the nonparametric regression of estimating an unknown d-variate function by using shallow ReLU neural networks. It is assumed that the regression function is from the Hölder space with smoothness α<(d+3)/2 or a variation space corresponding to shallow neural networks, which can be viewed as an infinitely wide neural network. In this setting, we prove that least squares estimators based on shallow neural networks with certain norm constraints on the weights are minimax optimal, if the network width is sufficiently large. As a byproduct, we derive a new size-independent bound for the local Rademacher complexity of shallow ReLU neural networks, which may be of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2023

Optimal rates of approximation by shallow ReLU^k neural networks and applications to nonparametric regression

We study the approximation capacity of some variation spaces correspondi...
research
09/18/2021

Near-Minimax Optimal Estimation With Shallow ReLU Neural Networks

We study the problem of estimating an unknown function from noisy data u...
research
08/17/2022

Shallow neural network representation of polynomials

We show that d-variate polynomials of degree R can be represented on [0,...
research
07/17/2022

Nonparametric regression with modified ReLU networks

We consider regression estimation with modified ReLU neural networks in ...
research
06/26/2022

Bounding the Width of Neural Networks via Coupled Initialization – A Worst Case Analysis

A common method in training neural networks is to initialize all the wei...
research
06/09/2021

Harmless Overparametrization in Two-layer Neural Networks

Overparametrized neural networks, where the number of active parameters ...

Please sign up or login with your details

Forgot password? Click here to reset