Achieve the Minimum Width of Neural Networks for Universal Approximation

09/23/2022
by   Yongqiang Cai, et al.
0

The universal approximation property (UAP) of neural networks is fundamental for deep learning, and it is well known that wide neural networks are universal approximators of continuous functions within both the L^p norm and the continuous/uniform norm. However, the exact minimum width, w_min, for the UAP has not been studied thoroughly. Recently, using a decoder-memorizer-encoder scheme, <cit.> found that w_min = max(d_x+1,d_y) for both the L^p-UAP of ReLU networks and the C-UAP of ReLU+STEP networks, where d_x,d_y are the input and output dimensions, respectively. In this paper, we consider neural networks with an arbitrary set of activation functions. We prove that both C-UAP and L^p-UAP for functions on compact domains share a universal lower bound of the minimal width; that is, w^*_min = max(d_x,d_y). In particular, the critical width, w^*_min, for L^p-UAP can be achieved by leaky-ReLU networks, provided that the input or output dimension is larger than one. Our construction is based on the approximation power of neural ordinary differential equations and the ability to approximate flow maps by neural networks. The nonmonotone or discontinuous activation functions case and the one-dimensional case are also discussed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/19/2023

Minimum width for universal approximation using ReLU networks on compact domain

The universal approximation property of width-bounded networks has been ...
research
11/10/2020

Expressiveness of Neural Networks Having Width Equal or Below the Input Dimension

The expressiveness of deep neural networks of bounded width has recently...
research
06/16/2020

Minimum Width for Universal Approximation

The universal approximation property of width-bounded networks has been ...
research
06/21/2019

Universal Approximation of Input-Output Maps by Temporal Convolutional Nets

There has been a recent shift in sequence-to-sequence modeling from recu...
research
05/15/2021

Universality and Optimality of Structured Deep Kernel Networks

Kernel based methods yield approximation models that are flexible, effic...
research
07/16/2022

Approximation Capabilities of Neural Networks using Morphological Perceptrons and Generalizations

Standard artificial neural networks (ANNs) use sum-product or multiply-a...
research
08/30/2023

Minimum Width for Deep, Narrow MLP: A Diffeomorphism and the Whitney Embedding Theorem Approach

Recently, there has been significant attention on determining the minimu...

Please sign up or login with your details

Forgot password? Click here to reset