Optimal Approximation Rates and Metric Entropy of ReLU^k and Cosine Networks

01/29/2021
by   Jonathan W. Siegel, et al.
1

This article addresses several fundamental issues associated with the approximation theory of neural networks, including the characterization of approximation spaces, the determination of the metric entropy of these spaces, and approximation rates of neural networks. For any activation function σ, we show that the largest Banach space of functions which can be efficiently approximated by the corresponding shallow neural networks is the space whose norm is given by the gauge of the closed convex hull of the set {±σ(ω· x + b)}. We characterize this space for the ReLU^k and cosine activation functions and, in particular, show that the resulting gauge space is equivalent to the spectral Barron space if σ=cos and is equivalent to the Barron space when σ= ReLU. Our main result establishes the precise asymptotics of the L^2-metric entropy of the unit ball of these guage spaces and, as a consequence, the optimal approximation rates for shallow ReLU^k networks. The sharpest previous results hold only in the special case that k=0 and d=2, where the metric entropy has been determined up to logarithmic factors. When k > 0 or d > 2, there is a significant gap between the previous best upper and lower bounds. We close all of these gaps and determine the precise asymptotics of the metric entropy for all k ≥ 0 and d≥ 2, including removing the logarithmic factors previously mentioned. Finally, we use these results to quantify how much is lost by Barron's spectral condition relative to the convex hull of {±σ(ω· x + b)} when σ= ReLU^k.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/28/2021

Sharp Lower Bounds on the Approximation Rate of Shallow Neural Networks

We consider the approximation rates of shallow neural networks with resp...
research
07/28/2023

Optimal Approximation of Zonoids and Uniform Approximation by Shallow Neural Networks

We study the following two related problems. The first is to determine t...
research
01/30/2023

Optimal Approximation Complexity of High-Dimensional Functions with Neural Networks

We investigate properties of neural networks that use both ReLU and x^2 ...
research
06/28/2021

Characterization of the Variation Spaces Corresponding to Shallow Neural Networks

We consider the variation space corresponding to a dictionary of functio...
research
03/06/2023

On the existence of optimal shallow feedforward networks with ReLU activation

We prove existence of global minima in the loss landscape for the approx...
research
04/05/2021

Deep neural network approximation of analytic functions

We provide an entropy bound for the spaces of neural networks with piece...
research
07/08/2020

Approximation with Neural Networks in Variable Lebesgue Spaces

This paper concerns the universal approximation property with neural net...

Please sign up or login with your details

Forgot password? Click here to reset