The training response law explains how deep neural networks learn

04/15/2022
by   Kenichi Nakazato, et al.
0

Deep neural network is the widely applied technology in this decade. In spite of the fruitful applications, the mechanism behind that is still to be elucidated. We study the learning process with a very simple supervised learning encoding problem. As a result, we found a simple law, in the training response, which describes neural tangent kernel. The response consists of a power law like decay multiplied by a simple response kernel. We can construct a simple mean-field dynamical model with the law, which explains how the network learns. In the learning, the input space is split into sub-spaces along competition between the kernels. With the iterated splits and the aging, the network gets more complexity, but finally loses its plasticity.

READ FULL TEXT
research
05/04/2018

Power Law in Sparsified Deep Neural Networks

The power law has been observed in the degree distributions of many biol...
research
10/23/2021

Learning curves for Gaussian process regression with power-law priors and targets

We study the power-law asymptotics of learning curves for Gaussian proce...
research
10/29/2010

Fractionally Predictive Spiking Neurons

Recent experimental work has suggested that the neural firing rate can b...
research
01/09/2019

Three Other Models of Computer System Performance

This note argues for more use of simple models beyond Amdahl's Law: Bott...
research
10/31/2022

A Law of Data Separation in Deep Learning

Multilayer neural networks have achieved superhuman performance in many ...
research
04/21/2022

Accelerating Machine Learning via the Weber-Fechner Law

The Weber-Fechner Law observes that human perception scales as the logar...
research
09/14/2023

The kernel-balanced equation for deep neural networks

Deep neural networks have shown many fruitful applications in this decad...

Please sign up or login with your details

Forgot password? Click here to reset