Why Deep Neural Networks for Function Approximation?

10/13/2016
by   Shiyu Liang, et al.
0

Recently there has been much interest in understanding why deep neural networks are preferred to shallow networks. We show that, for a large class of piecewise smooth functions, the number of neurons needed by a shallow network to approximate a function is exponentially larger than the corresponding number of neurons needed by a deep network for a given degree of function approximation. First, we consider univariate functions on a bounded interval and require a neural network to achieve an approximation error of ε uniformly over the interval. We show that shallow networks (i.e., networks whose depth does not depend on ε) require Ω(poly(1/ε)) neurons while deep networks (i.e., networks whose depth grows with 1/ε) require O(polylog(1/ε)) neurons. We then extend these results to certain classes of important multivariate functions. Our results are derived for neural networks which use a combination of rectifier linear units (ReLUs) and binary step units, two of the most popular type of activation functions. Our analysis builds on a simple observation: the multiplication of two bits can be represented by a ReLU.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/24/2023

Rates of Approximation by ReLU Shallow Neural Networks

Neural networks activated by the rectified linear unit (ReLU) play a cen...
research
02/01/2020

A Corrective View of Neural Networks: Representation, Memorization and Learning

We develop a corrective mechanism for neural network approximation: the ...
research
09/06/2018

Applying Deep Learning to Derivatives Valuation

The universal approximation theorem of artificial neural networks states...
research
01/11/2023

Exploring the Approximation Capabilities of Multiplicative Neural Networks for Smooth Functions

Multiplication layers are a key component in various influential neural ...
research
08/29/2016

Why does deep and cheap learning work so well?

We show how the success of deep learning could depend not only on mathem...
research
08/17/2022

Shallow neural network representation of polynomials

We show that d-variate polynomials of degree R can be represented on [0,...
research
05/03/2015

Some Theoretical Properties of a Network of Discretely Firing Neurons

The problem of optimising a network of discretely firing neurons is addr...

Please sign up or login with your details

Forgot password? Click here to reset