Optimal Approximation Rate of ReLU Networks in terms of Width and Depth

02/28/2021
by   Zuowei Shen, et al.
0

This paper concentrates on the approximation power of deep feed-forward neural networks in terms of width and depth. It is proved by construction that ReLU networks with width 𝒪(max{d⌊ N^1/d⌋, N+2}) and depth 𝒪(L) can approximate a Hölder continuous function on [0,1]^d with an approximation rate 𝒪(λ√(d) (N^2L^2ln N)^-α/d), where α∈ (0,1] and λ>0 are Hölder order and constant, respectively. Such a rate is optimal up to a constant in terms of width and depth separately, while existing results are only nearly optimal without the logarithmic factor in the approximation rate. More generally, for an arbitrary continuous function f on [0,1]^d, the approximation rate becomes 𝒪( √(d) ω_f( (N^2L^2ln N)^-1/d) ), where ω_f(·) is the modulus of continuity. We also extend our analysis to any continuous function f on a bounded set. Particularly, if ReLU networks with depth 31 and width 𝒪(N) are used to approximate one-dimensional Lipschitz continuous functions on [0,1] with a Lipschitz constant λ>0, the approximation rate in terms of the total number of parameters, W=𝒪(N^2), becomes 𝒪(λWln W), which has not been discovered in the literature for fixed-depth ReLU networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2020

Deep Network Approximation with Discrepancy Being Reciprocal of Width to Power of Depth

A new network with super approximation power is introduced. This network...
research
06/13/2019

Deep Network Approximation Characterized by Number of Neurons

This paper quantitatively characterizes the approximation power of deep ...
research
11/15/2021

ReLU Network Approximation in Terms of Intrinsic Parameters

This paper studies the approximation error of ReLU networks in terms of ...
research
05/19/2022

Neural Network Architecture Beyond Width and Depth

This paper proposes a new neural network architecture by introducing an ...
research
08/09/2017

Universal Function Approximation by Deep Neural Nets with Bounded Width and ReLU Activations

This article concerns the expressive power of depth in neural nets with ...
research
10/31/2017

Approximating Continuous Functions by ReLU Nets of Minimal Width

This article concerns the expressive power of depth in deep feed-forward...
research
02/28/2021

Deep Neural Networks with ReLU-Sine-Exponential Activations Break Curse of Dimensionality on Hölder Class

In this paper, we construct neural networks with ReLU, sine and 2^x as a...

Please sign up or login with your details

Forgot password? Click here to reset