PCA of high dimensional random walks with comparison to neural network training

06/22/2018
by   Joseph M. Antognini, et al.
0

One technique to visualize the training of neural networks is to perform PCA on the parameters over the course of training and to project to the subspace spanned by the first few PCA components. In this paper we compare this technique to the PCA of a high dimensional random walk. We compute the eigenvalues and eigenvectors of the covariance of the trajectory and prove that in the long trajectory and high dimensional limit most of the variance is in the first few PCA components, and that the projection of the trajectory onto any subspace spanned by PCA components is a Lissajous curve. We generalize these results to a random walk with momentum and to an Ornstein-Uhlenbeck processes (i.e., a random walk in a quadratic potential) and show that in high dimensions the walk is not mean reverting, but will instead be trapped at a fixed distance from the minimum. We finally compare the distribution of PCA variances and the PCA projected training trajectories of a linear model trained on CIFAR-10 and ResNet-50-v2 trained on Imagenet and find that the distribution of PCA variances resembles a random walk with drift.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2019

On the center of mass of the elephant random walk

Our goal is to investigate the asymptotic behavior of the center of mass...
research
07/18/2023

Pseudorandomness of the Sticky Random Walk

We extend the pseudorandomness of random walks on expander graphs using ...
research
04/23/2022

Discovering Intrinsic Reward with Contrastive Random Walk

The aim of this paper is to demonstrate the efficacy of using Contrastiv...
research
04/14/2021

A Gaussian fixed point random walk

In this note, we design a discrete random walk on the real line which ta...
research
11/27/2018

Knots in random neural networks

The weights of a neural network are typically initialized at random, and...
research
09/02/2019

Agreement testing theorems on layered set systems

We introduce a framework of layered subsets, and give a sufficient condi...
research
01/09/2018

Compressing Deep Neural Networks: A New Hashing Pipeline Using Kac's Random Walk Matrices

The popularity of deep learning is increasing by the day. However, despi...

Please sign up or login with your details

Forgot password? Click here to reset