Lightlike Neuromanifolds, Occam's Razor and Deep Learning

05/27/2019
by   Ke Sun, et al.
3

Why do deep neural networks generalize with a very high dimensional parameter space? We took an information theoretic approach. We find that the dimensionality of the parameter space can be studied by singular semi-Riemannian geometry and is upper-bounded by the sample size. We adapt Fisher information to this singular neuromanifold. We use random matrix theory to derive a minimum description length of a deep learning model, where the spectrum of the Fisher information matrix plays a key role to improve generalisation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/06/2020

On the minmax regret for statistical manifolds: the role of curvature

Model complexity plays an essential role in its selection, namely, by ch...
research
06/14/2020

The Spectrum of Fisher Information of Deep Networks Achieving Dynamical Isometry

The Fisher information matrix (FIM) is fundamental for understanding the...
research
10/31/2019

Connecting exciton diffusion with surface roughness via deep learning

Exciton diffusion plays a vital role in the function of many organic sem...
research
06/17/2022

On the Influence of Enforcing Model Identifiability on Learning dynamics of Gaussian Mixture Models

A common way to learn and analyze statistical models is to consider oper...
research
10/14/2019

Pathological spectra of the Fisher information metric and its variants in deep neural networks

The Fisher information matrix (FIM) plays an essential role in statistic...
research
12/07/2021

Towards Modeling and Resolving Singular Parameter Spaces using Stratifolds

When analyzing parametric statistical models, a useful approach consists...
research
06/10/2022

Fisher SAM: Information Geometry and Sharpness Aware Minimisation

Recent sharpness-aware minimisation (SAM) is known to find flat minima w...

Please sign up or login with your details

Forgot password? Click here to reset