On the Expressive Power of Overlapping Architectures of Deep Learning

03/06/2017
by   Or Sharir, et al.
0

Expressive efficiency refers to the relation between two architectures A and B, whereby any function realized by B could be replicated by A, but there exists functions realized by A, which cannot be replicated by B unless its size grows significantly larger. For example, it is known that deep networks are exponentially efficient with respect to shallow networks, in the sense that a shallow network must grow exponentially large in order to approximate the functions represented by a deep network of polynomial size. In this work, we extend the study of expressive efficiency to the attribute of network connectivity and in particular to the effect of "overlaps" in the convolutional process, i.e., when the stride of the convolution is smaller than its filter size (receptive field). Our analysis shows that having overlapping local receptive fields, and more broadly denser connectivity, results in an exponential increase in the expressive capacity of neural networks. Moreover, while denser connectivity can increase the expressive capacity, we show that the most common types of modern architectures already exhibit exponential increase in expressivity, without relying on fully-connected layers.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/20/2017

Boosting Dilated Convolutional Networks with Mixed Tensor Decompositions

The driving force behind deep networks is their ability to compactly rep...
research
01/29/2023

On Enhancing Expressive Power via Compositions of Single Fixed-Size ReLU Network

This paper studies the expressive power of deep neural networks from the...
research
01/02/2019

The capacity of feedforward neural networks

A long standing open problem in the theory of neural networks is the dev...
research
08/29/2022

Rosenblatt's first theorem and frugality of deep learning

First Rosenblatt's theorem about omnipotence of shallow networks states ...
research
11/21/2015

GradNets: Dynamic Interpolation Between Neural Architectures

In machine learning, there is a fundamental trade-off between ease of op...
research
09/16/2015

On the Expressive Power of Deep Learning: A Tensor Analysis

It has long been conjectured that hypotheses spaces suitable for data th...
research
06/11/2019

Analysis of Memory Capacity for Deep Echo State Networks

In this paper, the echo state network (ESN) memory capacity, which repre...

Please sign up or login with your details

Forgot password? Click here to reset