The power of deeper networks for expressing natural functions

05/16/2017
by   David Rolnick, et al.
0

It is well-known that neural networks are universal approximators, but that deeper networks tend to be much more efficient than shallow ones. We shed light on this by proving that the total number of neurons m required to approximate natural classes of multivariate polynomials of n variables grows only linearly with n for deep neural networks, but grows exponentially when merely a single hidden layer is allowed. We also provide evidence that when the number of hidden layers is increased from 1 to k, the neuron requirement grows exponentially not with n but with n^1/k, suggesting that the minimum number of layers required for computational tractability grows only logarithmically with n.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset