Correlation between entropy and generalizability in a neural network

07/05/2022
by   Ge Zhang, et al.
0

Although neural networks can solve very complex machine-learning problems, the theoretical reason for their generalizability is still not fully understood. Here we use Wang-Landau Mote Carlo algorithm to calculate the entropy (logarithm of the volume of a part of the parameter space) at a given test accuracy, and a given training loss function value or training accuracy. Our results show that entropical forces help generalizability. Although our study is on a very simple application of neural networks (a spiral dataset and a small, fully-connected neural network), our approach should be useful in explaining the generalizability of more complicated neural networks in future works.

READ FULL TEXT
research
01/22/2021

Sobolev Training for the Neural Network Solutions of PDEs

Approximating the numerical solutions of Partial Differential Equations ...
research
06/02/2020

Deep Learning in Target Space

Deep learning uses neural networks which are parameterised by their weig...
research
08/28/2023

Entropy-based Guidance of Deep Neural Networks for Accelerated Convergence and Improved Performance

Neural networks have dramatically increased our capacity to learn from l...
research
02/04/2021

Universal Approximation Theorems of Fully Connected Binarized Neural Networks

Neural networks (NNs) are known for their high predictive accuracy in co...
research
07/03/2021

Slope and generalization properties of neural networks

Neural networks are very successful tools in for example advanced classi...
research
10/04/2018

Machine learning electron correlation in a disordered medium

Learning from data has led to a paradigm shift in computational material...
research
05/26/2020

Machine Learning-Based Unbalance Detection of a Rotating Shaft Using Vibration Data

Fault detection at rotating machinery with the help of vibration sensors...

Please sign up or login with your details

Forgot password? Click here to reset