Stochastic Gradient Descent-induced drift of representation in a two-layer neural network

02/06/2023
by   Farhad Pashakhanloo, et al.
0

Representational drift refers to over-time changes in neural activation accompanied by a stable task performance. Despite being observed in the brain and in artificial networks, the mechanisms of drift and its implications are not fully understood. Motivated by recent experimental findings of stimulus-dependent drift in the piriform cortex, we use theory and simulations to study this phenomenon in a two-layer linear feedforward network. Specifically, in a continual learning scenario, we study the drift induced by the noise inherent in the Stochastic Gradient Descent (SGD). By decomposing the learning dynamics into the normal and tangent spaces of the minimum-loss manifold, we show the former correspond to a finite variance fluctuation, while the latter could be considered as an effective diffusion process on the manifold. We analytically compute the fluctuation and the diffusion coefficients for the stimuli representations in the hidden layer as a function of network parameters and input distribution. Further, consistent with experiments, we show that the drift rate is slower for a more frequently presented stimulus. Overall, our analysis yields a theoretical framework for better understanding of the drift phenomenon in biological and artificial neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/10/2023

Is Learning in Biological Neural Networks based on Stochastic Gradient Descent? An analysis using stochastic processes

In recent years, there has been an intense debate about how learning in ...
research
08/24/2020

Noise-induced degeneration in online learning

In order to elucidate the plateau phenomena caused by vanishing gradient...
research
07/11/2022

On the Stochastic Gradient Descent and Inverse Variance-flatness Relation in Artificial Neural Networks

Stochastic gradient descent (SGD), a widely used algorithm in deep-learn...
research
05/22/2017

Batch Size Matters: A Diffusion Approximation Framework on Nonconvex Stochastic Gradient Descent

In this paper, we study the stochastic gradient descent method in analyz...
research
06/02/2021

Random walk approximation for irreversible drift-diffusion process on manifold: ergodicity, unconditional stability and convergence

Irreversible drift-diffusion processes are very common in biochemical re...
research
12/04/2019

A probability theoretic approach to drifting data in continuous time domains

The notion of drift refers to the phenomenon that the distribution, whic...

Please sign up or login with your details

Forgot password? Click here to reset