Best k-layer neural network approximations

07/02/2019
by   Lek-Heng Lim, et al.
0

We investigate the geometry of the empirical risk minimization problem for k-layer neural networks. We will provide examples showing that for the classical activation functions σ(x)= 1/(1 + (-x)) and σ(x)=(x), there exists a positive-measured subset of target functions that do not have best approximations by a fixed number of layers of neural networks. In addition, we study in detail the properties of shallow networks, classifying cases when a best k-layer neural network approximation always exists or does not exist for the ReLU activation σ=(0,x). We also determine the dimensions of shallow ReLU-activated networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/03/2016

Error bounds for approximations with deep ReLU networks

We study expressive power of shallow and deep neural networks with piece...
research
03/06/2023

On the existence of optimal shallow feedforward networks with ReLU activation

We prove existence of global minima in the loss landscape for the approx...
research
07/03/2023

Neural Polytopes

We find that simple neural networks with ReLU activation generate polyto...
research
01/17/2020

Approximating Activation Functions

ReLU is widely seen as the default choice for activation functions in ne...
research
05/27/2019

Equivalent and Approximate Transformations of Deep Neural Networks

Two networks are equivalent if they produce the same output for any give...
research
12/17/2020

Reduced Order Modeling using Shallow ReLU Networks with Grassmann Layers

This paper presents a nonlinear model reduction method for systems of eq...
research
04/06/2018

A comparison of deep networks with ReLU activation function and linear spline-type methods

Deep neural networks (DNNs) generate much richer function spaces than sh...

Please sign up or login with your details

Forgot password? Click here to reset