On the number of response regions of deep feed forward networks with piece-wise linear activations

12/20/2013
by   Razvan Pascanu, et al.
0

This paper explores the complexity of deep feedforward networks with linear pre-synaptic couplings and rectified linear activations. This is a contribution to the growing body of work contrasting the representational power of deep and shallow network architectures. In particular, we offer a framework for comparing deep and shallow models that belong to the family of piecewise linear functions based on computational geometry. We look at a deep rectifier multi-layer perceptron (MLP) with linear outputs units and compare it with a single layer version of the model. In the asymptotic regime, when the number of inputs stays constant, if the shallow model has kn hidden units and n_0 inputs, then the number of linear regions is O(k^n_0n^n_0). For a k layer model with n hidden units on each layer it is Ω(n/n_0^k-1n^n_0). The number n/n_0^k-1 grows faster than k^n_0 when n tends to infinity or when k tends to infinity and n ≥ 2n_0. Additionally, even when k is small, if we restrict n to be 2n_0, we can show that a deep model has considerably more linear regions that a shallow one. We consider this as a first step towards understanding the complexity of these models and specifically towards providing suitable mathematical tools for future analysis.

READ FULL TEXT

page 8

page 13

research
02/08/2014

On the Number of Linear Regions of Deep Neural Networks

We study the complexity of functions computable by deep feedforward neur...
research
08/11/2023

Automated Sizing and Training of Efficient Deep Autoencoders using Second Order Algorithms

We propose a multi-step training method for designing generalized linear...
research
05/22/2018

A Tropical Approach to Neural Networks with Piecewise Linear Activations

We present a new, unifying approach following some recent developments o...
research
08/29/2022

Rosenblatt's first theorem and frugality of deep learning

First Rosenblatt's theorem about omnipotence of shallow networks states ...
research
05/27/2019

Equivalent and Approximate Transformations of Deep Neural Networks

Two networks are equivalent if they produce the same output for any give...
research
11/20/2020

Normalization effects on shallow neural networks and related asymptotic expansions

We consider shallow (single hidden layer) neural networks and characteri...
research
03/22/2022

Clustering units in neural networks: upstream vs downstream information

It has been hypothesized that some form of "modular" structure in artifi...

Please sign up or login with your details

Forgot password? Click here to reset