Lower bounds over Boolean inputs for deep neural networks with ReLU gates

11/08/2017
by   Anirbit Mukherjee, et al.
0

Motivated by the resurgence of neural networks in being able to solve complex learning tasks we undertake a study of high depth networks using ReLU gates which implement the function x {0,x}. We try to understand the role of depth in such neural networks by showing size lowerbounds against such network architectures in parameter regimes hitherto unexplored. In particular we show the following two main results about neural nets computing Boolean functions of input dimension n, 1. We use the method of random restrictions to show almost linear, Ω(ϵ^2(1-δ)n^1-δ), lower bound for completely weight unrestricted LTF-of-ReLU circuits to match the Andreev function on at least 1/2 +ϵ fraction of the inputs for ϵ > √(2^2/2-δ(n)/n) for any δ∈ (0,1/2) 2. We use the method of sign-rank to show exponential in dimension lower bounds for ReLU circuits ending in a LTF gate and of depths upto O(n^ξ) with ξ < 1/8 with some restrictions on the weights in the bottom most layer. All other weights in these circuits are kept unrestricted. This in turns also implies the same lowerbounds for LTF circuits with the same architecture and the same weight restrictions on their bottom most layer. Along the way we also show that there exists a R^ n→R Sum-of-ReLU-of-ReLU function which Sum-of-ReLU neural nets can never represent no matter how large they are allowed to be.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/05/2023

Top-Down Lower Bounds for Depth-Four Circuits

We present a top-down lower-bound method for depth-4 boolean circuits. I...
research
11/25/2019

Trajectory growth lower bounds for random sparse deep ReLU networks

This paper considers the growth in the length of one-dimensional traject...
research
05/18/2018

Two geometric input transformation methods for fast online reinforcement learning with neural nets

We apply neural nets with ReLU gates in online reinforcement learning. O...
research
02/14/2016

Benefits of depth in neural networks

For any positive integer k, there exist neural networks with Θ(k^3) laye...
research
10/25/2021

A Compilation of Succinctness Results for Arithmetic Circuits

Arithmetic circuits (AC) are circuits over the real numbers with 0/1-val...
research
05/31/2021

Towards Lower Bounds on the Depth of ReLU Neural Networks

We contribute to a better understanding of the class of functions that i...
research
10/02/2019

Identifying Weights and Architectures of Unknown ReLU Networks

The output of a neural network depends on its parameters in a highly non...

Please sign up or login with your details

Forgot password? Click here to reset