Neural Network Approximation: Three Hidden Layers Are Enough

10/25/2020
by   Zuowei Shen, et al.
0

A three-hidden-layer neural network with super approximation power is introduced. This network is built with the Floor function (⌊ x⌋), the exponential function (2^x), the step function (_x≥ 0), or their compositions as activation functions in each neuron and hence we call such networks as Floor-Exponential-Step (FLES) networks. For any width hyper-parameter N∈ℕ^+, it is shown that FLES networks with a width max{d, N} and three hidden layers can uniformly approximate a Hölder function f on [0,1]^d with an exponential approximation rate 3λ d^α/22^-α N, where α∈(0,1] and λ are the Hölder order and constant, respectively. More generally for an arbitrary continuous function f on [0,1]^d with a modulus of continuity ω_f(·), the constructive approximation rate is ω_f(√(d) 2^-N)+2ω_f(√(d))2^-N. As a consequence, this new class of networks overcomes the curse of dimensionality in approximation power when the variation of ω_f(r) as r→ 0 is moderate (e.g., ω_f(r)≲ r^α for Hölder continuous functions), since the major term to be concerned in our approximation rate is essentially √(d) times a function of N independent of d within the modulus of continuity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2020

Deep Network Approximation with Discrepancy Being Reciprocal of Width to Power of Depth

A new network with super approximation power is introduced. This network...
research
02/28/2021

Deep Neural Networks with ReLU-Sine-Exponential Activations Break Curse of Dimensionality on Hölder Class

In this paper, we construct neural networks with ReLU, sine and 2^x as a...
research
06/13/2019

Deep Network Approximation Characterized by Number of Neurons

This paper quantitatively characterizes the approximation power of deep ...
research
11/25/2022

LU decomposition and Toeplitz decomposition of a neural network

It is well-known that any matrix A has an LU decomposition. Less well-kn...
research
02/26/2019

Nonlinear Approximation via Compositions

We study the approximation efficiency of function compositions in nonlin...
research
06/28/2018

ResNet with one-neuron hidden layers is a Universal Approximator

We demonstrate that a very deep ResNet with stacked modules with one neu...
research
08/07/2023

Noncompact uniform universal approximation

The universal approximation theorem is generalised to uniform convergenc...

Please sign up or login with your details

Forgot password? Click here to reset