Finite Sample Identification of Wide Shallow Neural Networks with Biases

11/08/2022
by   Massimo Fornasier, et al.
0

Artificial neural networks are functions depending on a finite number of parameters typically encoded as weights and biases. The identification of the parameters of the network from finite samples of input-output pairs is often referred to as the teacher-student model, and this model has represented a popular framework for understanding training and generalization. Even if the problem is NP-complete in the worst case, a rapidly growing literature – after adding suitable distributional assumptions – has established finite sample identification of two-layer networks with a number of neurons m=𝒪(D), D being the input dimension. For the range D<m<D^2 the problem becomes harder, and truly little is known for networks parametrized by biases as well. This paper fills the gap by providing constructive methods and theoretical guarantees of finite sample identification for such wider shallow networks with biases. Our approach is based on a two-step pipeline: first, we recover the direction of the weights, by exploiting second order information; next, we identify the signs by suitable algebraic evaluations, and we recover the biases by empirical risk minimization via gradient descent. Numerical results demonstrate the effectiveness of our approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2021

Stable Recovery of Entangled Weights: Towards Robust Identification of Deep Neural Networks from Minimal Samples

In this paper we approach the problem of unique and stable identifiabili...
research
06/27/2020

Optimization and Generalization of Shallow Neural Networks with Quadratic Activation Functions

We study the dynamics of optimization and the generalization properties ...
research
06/30/2019

Robust and Resource Efficient Identification of Two Hidden Layer Neural Networks

We address the structure identification and the uniform approximation of...
research
09/17/2022

Approximation results for Gradient Descent trained Shallow Neural Networks in 1d

Two aspects of neural networks that have been extensively studied in the...
research
10/17/2018

Finite sample expressive power of small-width ReLU networks

We study universal finite sample expressivity of neural networks, define...
research
02/18/2021

Reduced-Order Neural Network Synthesis with Robustness Guarantees

In the wake of the explosive growth in smartphones and cyberphysical sys...
research
12/14/2020

System identification of biophysical neuronal models

After sixty years of quantitative biophysical modeling of neurons, the i...

Please sign up or login with your details

Forgot password? Click here to reset