Nonclosedness of the Set of Neural Networks in Sobolev Space

07/23/2020
by   Scott Mahan, et al.
0

We examine the closedness of the set of realized neural networks of a fixed architecture in Sobolev space. For an exactly m-times differentiable activation function ρ, we construct a sequence of neural networks (Φ_n)_n ∈ℕ whose realizations converge in order-(m-1) Sobolev norm to a function that cannot be realized exactly by a neural network. Thus, the set of realized neural networks is not closed in the order-(m-1) Sobolev space W^m-1,p. We further show that this set is not closed in W^m,p under slightly stronger conditions on the m-th derivative of ρ. For a real analytic activation function, we show that the set of realized neural networks is not closed in W^k,p for any k ∈ℕ. These results suggest that training a network to approximate a target function in Sobolev norm does not prevent parameter explosion. Finally, we present experimental results demonstrating that parameter explosion occurs in stochastic training regardless of the norm under which the network is trained. However, the network is still capable of closely approximating a non-network target function with network parameters that grow at a manageable rate.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/03/2016

A continuum among logarithmic, linear, and exponential functions, and its potential to improve generalization in neural networks

We present the soft exponential activation function for artificial neura...
research
10/07/2019

Neural network integral representations with the ReLU activation function

We derive a formula for neural network integral representations on the s...
research
01/11/2022

Deep Neural Network Approximation For Hölder Functions

In this work, we explore the approximation capability of deep Rectified ...
research
04/10/2023

Criticality versus uniformity in deep neural networks

Deep feedforward networks initialized along the edge of chaos exhibit ex...
research
03/22/2020

TanhExp: A Smooth Activation Function with High Convergence Speed for Lightweight Neural Networks

Lightweight or mobile neural networks used for real-time computer vision...
research
02/07/2023

Efficient Parametric Approximations of Neural Network Function Space Distance

It is often useful to compactly summarize important properties of model ...
research
03/11/2020

Improving Convolutional Neural Networks Via Conservative Field Regularisation and Integration

Current research in convolutional neural networks (CNN) focuses mainly o...

Please sign up or login with your details

Forgot password? Click here to reset