Towards a Phenomenological Understanding of Neural Networks: Data

05/01/2023
by   Samuel Tovey, et al.
2

A theory of neural networks (NNs) built upon collective variables would provide scientists with the tools to better understand the learning process at every stage. In this work, we introduce two such variables, the entropy and the trace of the empirical neural tangent kernel (NTK) built on the training data passed to the model. We empirically analyze the NN performance in the context of these variables and find that there exists correlation between the starting entropy, the trace of the NTK, and the generalization of the model computed after training is complete. This framework is then applied to the problem of optimal data selection for the training of NNs. To this end, random network distillation (RND) is used as a means of selecting training data which is then compared with random selection of data. It is shown that not only does RND select data-sets capable of outperforming random selection, but that the collective variables associated with the RND data-sets are larger than those of the randomly selected sets. The results of this investigation provide a stable ground from which the selection of data for NN training can be driven by this phenomenological framework.

READ FULL TEXT

page 7

page 12

page 13

research
05/28/2023

Repeated Random Sampling for Minimizing the Time-to-Accuracy of Learning

Methods for carefully selecting or generating a small set of training da...
research
05/28/2022

A Quadrature Perspective on Frequency Bias in Neural Network Training with Nonuniform Data

Small generalization errors of over-parameterized neural networks (NNs) ...
research
02/11/2020

A study of local optima for learning feature interactions using neural networks

In many fields such as bioinformatics, high energy physics, power distri...
research
10/26/2020

The estimation of training accuracy for two-layer neural networks on random datasets without training

Although the neural network (NN) technique plays an important role in ma...
research
05/17/2018

DNN or k-NN: That is the Generalize vs. Memorize Question

This paper studies the relationship between the classification performed...
research
01/20/2021

Can stable and accurate neural networks be computed? – On the barriers of deep learning and Smale's 18th problem

Deep learning (DL) has had unprecedented success and is now entering sci...
research
11/16/2018

nn-dependability-kit: Engineering Neural Networks for Safety-Critical Systems

nn-dependability-kit is an open-source toolbox to support safety enginee...

Please sign up or login with your details

Forgot password? Click here to reset