Gradient descent in Gaussian random fields as a toy model for high-dimensional optimisation in deep learning

03/24/2018
by   Mariano Chouza, et al.
0

In this paper we model the loss function of high-dimensional optimization problems by a Gaussian random field, or equivalently a Gaussian process. Our aim is to study gradient descent in such loss functions or energy landscapes and compare it to results obtained from real high-dimensional optimization problems such as encountered in deep learning. In particular, we analyze the distribution of the improved loss function after a step of gradient descent, provide analytic expressions for the moments as well as prove asymptotic normality as the dimension of the parameter space becomes large. Moreover, we compare this with the expectation of the global minimum of the landscape obtained by means of the Euler characteristic of excursion sets. Besides complementing our analytical findings with numerical results from simulated Gaussian random fields, we also compare it to loss functions obtained from optimisation problems on synthetic and real data sets by proposing a "black box" random field toy-model for a deep neural network loss function.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2021

Combining resampling and reweighting for faithful stochastic optimization

Many machine learning and data science tasks require solving non-convex ...
research
02/07/2019

Combining learning rate decay and weight decay with complexity gradient descent - Part I

The role of L^2 regularization, in the specific case of deep neural netw...
research
07/15/2021

SA-GD: Improved Gradient Descent Learning Strategy with Simulated Annealing

Gradient descent algorithm is the most utilized method when optimizing m...
research
06/21/2022

A consistent and flexible framework for deep matrix factorizations

Deep matrix factorizations (deep MFs) are recent unsupervised data minin...
research
01/22/2018

Rover Descent: Learning to optimize by learning to navigate on prototypical loss surfaces

Learning to optimize - the idea that we can learn from data algorithms t...
research
03/26/2021

UMAP does not reproduce high-dimensional similarities due to negative sampling

UMAP has supplanted t-SNE as state-of-the-art for visualizing high-dimen...
research
08/25/2021

Excess and Deficiency of Extreme Multidimensional Random Fields

Probability distributions and densities are derived for the excess and d...

Please sign up or login with your details

Forgot password? Click here to reset