A free from local minima algorithm for training regressive MLP neural networks

08/22/2023
by   Augusto Montisci, et al.
0

In this article an innovative method for training regressive MLP networks is presented, which is not subject to local minima. The Error-Back-Propagation algorithm, proposed by William-Hinton-Rummelhart, has had the merit of favouring the development of machine learning techniques, which has permeated every branch of research and technology since the mid-1980s. This extraordinary success is largely due to the black-box approach, but this same factor was also seen as a limitation, as soon more challenging problems were approached. One of the most critical aspects of the training algorithms was that of local minima of the loss function, typically the mean squared error of the output on the training set. In fact, as the most popular training algorithms are driven by the derivatives of the loss function, there is no possibility to evaluate if a reached minimum is local or global. The algorithm presented in this paper avoids the problem of local minima, as the training is based on the properties of the distribution of the training set, or better on its image internal to the neural network. The performance of the algorithm is shown for a well-known benchmark.

READ FULL TEXT
research
04/20/2023

Interpolation property of shallow neural networks

We study the geometry of global minima of the loss landscape of overpara...
research
10/30/2018

Piecewise Strong Convexity of Neural Networks

We study the loss surface of a fully connected neural network with ReLU ...
research
04/01/2022

Estimating the Jacobian matrix of an unknown multivariate function from sample values by means of a neural network

We describe, implement and test a novel method for training neural netwo...
research
06/16/2021

Analysis and Optimisation of Bellman Residual Errors with Neural Function Approximation

Recent development of Deep Reinforcement Learning has demonstrated super...
research
11/12/2022

On the High Symmetry of Neural Network Functions

Training neural networks means solving a high-dimensional optimization p...
research
10/27/2020

Wide flat minima and optimal generalization in classifying high-dimensional Gaussian mixtures

We analyze the connection between minimizers with good generalizing prop...
research
06/26/2023

Black holes and the loss landscape in machine learning

Understanding the loss landscape is an important problem in machine lear...

Please sign up or login with your details

Forgot password? Click here to reset