Regularization, early-stopping and dreaming: a Hopfield-like setup to address generalization and overfitting

08/01/2023
by   Elena Agliari, et al.
0

In this work we approach attractor neural networks from a machine learning perspective: we look for optimal network parameters by applying a gradient descent over a regularized loss function. Within this framework, the optimal neuron-interaction matrices turn out to be a class of matrices which correspond to Hebbian kernels revised by iteratively applying some unlearning protocols. Remarkably, the number of unlearning steps is proved to be related to the regularization hyperparameters of the loss function and to the training time. Thus, we can design strategies to avoid overfitting that are formulated in terms of the algebraic properties of the interaction matrix, or, equivalently, in terms of regularization tuning and early-stopping strategies. The generalization capabilities of these attractor networks are also investigated: analytical results are obtained for random synthetic datasets, next, the emerging picture is corroborated by numerical experiments that highlight the existence of several regimes (i.e., overfitting, failure and success) as the dataset parameters are varied.

READ FULL TEXT
research
07/06/2020

Regularization Matters: A Nonparametric Perspective on Overparametrized Neural Network

Overparametrized neural networks trained by gradient descent (GD) can pr...
research
07/26/2018

A Collaborative Approach to Angel and Venture Capital Investment Recommendations

Matrix factorization was used to generate investment recommendations for...
research
04/16/2011

Adding noise to the input of a model trained with a regularized objective

Regularization is a well studied problem in the context of neural networ...
research
09/23/2021

Deep Learning with Kernel Flow Regularization for Time Series Forecasting

Long Short-Term Memory (LSTM) neural networks have been widely used for ...
research
02/13/2023

Calibrating a Deep Neural Network with Its Predecessors

Confidence calibration - the process to calibrate the output probability...
research
11/26/2021

Implicit Data-Driven Regularization in Deep Neural Networks under SGD

Much research effort has been devoted to explaining the success of deep ...
research
08/15/2015

A Comparative Study on Regularization Strategies for Embedding-based Neural Networks

This paper aims to compare different regularization strategies to addres...

Please sign up or login with your details

Forgot password? Click here to reset