Why does deep and cheap learning work so well?

08/29/2016
by   Henry W. Lin, et al.
0

We show how the success of deep learning could depend not only on mathematics but also on physics: although well-known mathematical theorems guarantee that neural networks can approximate arbitrary functions well, the class of functions of practical interest can frequently be approximated through "cheap learning" with exponentially fewer parameters than generic ones. We explore how properties frequently encountered in physics such as symmetry, locality, compositionality, and polynomial log-probability translate into exceptionally simple neural networks. We further argue that when the statistical process generating the data is of a certain hierarchical form prevalent in physics and machine-learning, a deep neural network can be more efficient than a shallow one. We formalize these claims using information theory and discuss the relation to the renormalization group. We prove various "no-flattening theorems" showing when efficient linear deep networks cannot be accurately approximated by shallow ones without efficiency loss, for example, we show that n variables cannot be multiplied using fewer than 2^n neurons in a single hidden layer.

READ FULL TEXT

page 2

page 8

research
02/15/2019

Efficient Deep Learning of GMMs

We show that a collection of Gaussian mixture models (GMMs) in R^n can b...
research
10/13/2016

Why Deep Neural Networks for Function Approximation?

Recently there has been much interest in understanding why deep neural n...
research
03/08/2019

Is Deeper Better only when Shallow is Good?

Understanding the power of depth in feed-forward neural networks is an o...
research
06/16/2016

Exponential expressivity in deep neural networks through transient chaos

We combine Riemannian geometry with the mean field theory of high dimens...
research
07/19/2019

Representational Capacity of Deep Neural Networks -- A Computing Study

There is some theoretical evidence that deep neural networks with multip...
research
05/16/2017

The power of deeper networks for expressing natural functions

It is well-known that neural networks are universal approximators, but t...
research
08/14/2016

Stacked Approximated Regression Machine: A Simple Deep Learning Approach

With the agreement of my coauthors, I Zhangyang Wang would like to withd...

Please sign up or login with your details

Forgot password? Click here to reset