Gradient Descent on Infinitely Wide Neural Networks: Global Convergence and Generalization

10/15/2021
by   Francis Bach, et al.
0

Many supervised machine learning methods are naturally cast as optimization problems. For prediction models which are linear in their parameters, this often leads to convex problems for which many mathematical guarantees exist. Models which are non-linear in their parameters such as neural networks lead to non-convex optimization problems for which guarantees are harder to obtain. In this review paper, we consider two-layer neural networks with homogeneous activation functions where the number of hidden neurons tends to infinity, and show how qualitative convergence guarantees may be derived.

READ FULL TEXT

page 12

page 17

research
03/31/2021

CDiNN -Convex Difference Neural Networks

Neural networks with ReLU activation function have been shown to be univ...
research
06/10/2017

Recovery Guarantees for One-hidden-layer Neural Networks

In this paper, we consider regression problems with one-hidden-layer neu...
research
02/18/2018

Neural Networks with Finite Intrinsic Dimension have no Spurious Valleys

Neural networks provide a rich class of high-dimensional, non-convex opt...
research
05/17/2019

SSFN: Self Size-estimating Feed-forward Network and Low Complexity Design

We design a self size-estimating feed-forward network (SSFN) using a joi...
research
08/26/2022

Non-probabilistic Supervised Learning for Non-linear Convex Variational Problems

In this article we propose, based on a non-probabilistic supervised lear...
research
03/06/2019

Why Learning of Large-Scale Neural Networks Behaves Like Convex Optimization

In this paper, we present some theoretical work to explain why simple gr...
research
10/10/2020

Maximin Optimization for Binary Regression

We consider regression problems with binary weights. Such optimization p...

Please sign up or login with your details

Forgot password? Click here to reset