A general approximation lower bound in L^p norm, with applications to feed-forward neural networks

06/09/2022
by   El Mehdi Achour, et al.
0

We study the fundamental limits to the expressive power of neural networks. Given two sets F, G of real-valued functions, we first prove a general lower bound on how well functions in F can be approximated in L^p(μ) norm by functions in G, for any p ≥ 1 and any probability measure μ. The lower bound depends on the packing number of F, the range of F, and the fat-shattering dimension of G. We then instantiate this bound to the case where G corresponds to a piecewise-polynomial feed-forward neural network, and describe in details the application to two sets F: Hölder balls and multivariate monotonic functions. Beside matching (known or new) upper bounds up to log factors, our lower bounds shed some light on the similarities or differences between approximation in L^p norm or in sup norm, solving an open question by DeVore et al. (2021). Our proof strategy differs from the sup norm case and uses a key probability result of Mendelson (2002).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2015

Norm-Based Capacity Control in Neural Networks

We investigate the capacity, convexity and characterization of a general...
research
01/24/2022

Approximation bounds for norm constrained neural networks with applications to regression and GANs

This paper studies the approximation capacity of ReLU neural networks wi...
research
04/22/2008

Natural pseudo-distance and optimal matching between reduced size functions

This paper studies the properties of a new lower bound for the natural p...
research
02/20/2022

Tight Bounds for Sketching the Operator Norm, Schatten Norms, and Subspace Embeddings

We consider the following oblivious sketching problem: given ϵ∈ (0,1/3) ...
research
04/04/2018

Identification of Shallow Neural Networks by Fewest Samples

We address the uniform approximation of sums of ridge functions ∑_i=1^m ...
research
04/28/2020

Adversarial Learning Guarantees for Linear Hypotheses and Neural Networks

Adversarial or test time robustness measures the susceptibility of a cla...
research
10/10/2018

The Fundamental Theorem of Algebra in ACL2

We report on a verification of the Fundamental Theorem of Algebra in ACL...

Please sign up or login with your details

Forgot password? Click here to reset