On the Expressive Power of Deep Learning: A Tensor Analysis

09/16/2015
by   Nadav Cohen, et al.
0

It has long been conjectured that hypotheses spaces suitable for data that is compositional in nature, such as text or images, may be more efficiently represented with deep hierarchical networks than with shallow ones. Despite the vast empirical evidence supporting this belief, theoretical justifications to date are limited. In particular, they do not account for the locality, sharing and pooling constructs of convolutional networks, the most successful deep learning architecture to date. In this work we derive a deep network architecture based on arithmetic circuits that inherently employs locality, sharing and pooling. An equivalence between the networks and hierarchical tensor factorizations is established. We show that a shallow network corresponds to CP (rank-1) decomposition, whereas a deep network corresponds to Hierarchical Tucker decomposition. Using tools from measure theory and matrix algebra, we prove that besides a negligible set, all functions that can be implemented by a deep network of polynomial size, require exponential size in order to be realized (or even approximated) by a shallow network. Since log-space computation transforms our networks into SimNets, the result applies directly to a deep learning architecture demonstrating promising empirical performance. The construction and theory developed in this paper shed new light on various practices and ideas employed by the deep learning community.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2016

Convolutional Rectifier Networks as Generalized Tensor Decompositions

Convolutional rectifier networks, i.e. convolutional neural networks wit...
research
05/05/2017

Analysis and Design of Convolutional Networks via Hierarchical Tensor Decompositions

The driving force behind convolutional networks - the most successful de...
research
05/23/2019

Tucker Decomposition Network: Expressive Power and Comparison

Deep neural networks have achieved a great success in solving many machi...
research
05/22/2016

Inductive Bias of Deep Convolutional Networks through Pooling Geometry

Our formal understanding of the inductive bias that drives the success o...
research
03/06/2017

On the Expressive Power of Overlapping Architectures of Deep Learning

Expressive efficiency refers to the relation between two architectures A...
research
03/20/2017

Boosting Dilated Convolutional Networks with Mixed Tensor Decompositions

The driving force behind deep networks is their ability to compactly rep...
research
04/05/2017

Deep Learning and Quantum Entanglement: Fundamental Connections with Implications to Network Design

Deep convolutional networks have witnessed unprecedented success in vari...

Please sign up or login with your details

Forgot password? Click here to reset