Universal Approximation with Deep Narrow Networks

05/21/2019
by   Patrick Kidger, et al.
5

The classical Universal Approximation Theorem certifies that the universal approximation property holds for the class of neural networks of arbitrary width. Here we consider the natural `dual' theorem for width-bounded networks of arbitrary depth. Precisely, let n be the number of inputs neurons, m be the number of output neurons, and let ρ be any nonaffine continuous function, with a continuous nonzero derivative at some point. Then we show that the class of neural networks of arbitrary depth, width n + m + 2, and activation function ρ, exhibits the universal approximation property with respect to the uniform norm on compact subsets of R^n. This covers every activation function possible to use in practice; in particular this includes polynomial activation functions, making this genuinely different to the classical case. We go on to establish some natural extensions of this result. Firstly, we show an analogous result for a certain class of nowhere differentiable activation functions. Secondly, we establish an analogous result for noncompact domains, by showing that deep narrow networks with the ReLU activation function exhibit the universal approximation property with respect to the p-norm on R^n. Finally, we show that width of only n + m + 1 suffices for `most' activation functions (whilst it is known that width of n + m - 1 does not suffice in general).

READ FULL TEXT
research
05/26/2023

Universal approximation with complex-valued deep narrow neural networks

We study the universality of complex-valued neural networks with bounded...
research
09/23/2021

Arbitrary-Depth Universal Approximation Theorems for Operator Neural Networks

The standard Universal Approximation Theorem for operator neural network...
research
07/06/2021

Deep Network Approximation: Achieving Arbitrary Accuracy with Fixed Number of Neurons

This paper develops simple feed-forward neural networks that achieve the...
research
11/25/2022

LU decomposition and Toeplitz decomposition of a neural network

It is well-known that any matrix A has an LU decomposition. Less well-kn...
research
08/30/2023

Minimum Width for Deep, Narrow MLP: A Diffeomorphism and the Whitney Embedding Theorem Approach

Recently, there has been significant attention on determining the minimu...
research
05/14/2015

Neural Network with Unbounded Activation Functions is Universal Approximator

This paper presents an investigation of the approximation property of ne...
research
11/10/2020

Expressiveness of Neural Networks Having Width Equal or Below the Input Dimension

The expressiveness of deep neural networks of bounded width has recently...

Please sign up or login with your details

Forgot password? Click here to reset