A Fully Online Approach for Covariance Matrices Estimation of Stochastic Gradient Descent Solutions

02/10/2020
by   Wanrong Zhu, et al.
10

Stochastic gradient descent (SGD) algorithm is widely used for parameter estimation especially in online setting. While this recursive algorithm is popular for computation and memory efficiency, the problem of quantifying variability and randomness of the solutions has been rarely studied. This paper aims at conducting statistical inference of SGD-based estimates in online setting. In particular, we propose a fully online estimator for the covariance matrix of averaged SGD iterates (ASGD). Based on the classic asymptotic normality results of ASGD, we construct asymptotically valid confidence intervals for model parameters. Upon receiving new observations, we can quickly update the covariance estimator and confidence intervals. This approach fits in online setting even if the total number of data is unknown and takes the full advantage of SGD: efficiency in both computation and memory.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/27/2016

Statistical Inference for Model Parameters in Stochastic Gradient Descent

The stochastic gradient descent (SGD) algorithm has been widely used in ...
research
11/04/2019

Statistical Inference for Model Parameters in Stochastic Gradient Descent via Batch Means

Statistical inference of true model parameters based on stochastic gradi...
research
06/25/2022

Statistical inference with implicit SGD: proximal Robbins-Monro vs. Polyak-Ruppert

The implicit stochastic gradient descent (ISGD), a proximal version of S...
research
04/02/2021

A Sieve Stochastic Gradient Descent Estimator for Online Nonparametric Regression in Sobolev ellipsoids

The goal of regression is to recover an unknown underlying function that...
research
12/02/2022

Covariance Estimators for the ROOT-SGD Algorithm in Online Learning

Online learning naturally arises in many statistical and machine learnin...
research
04/03/2019

Normal Approximation for Stochastic Gradient Descent via Non-Asymptotic Rates of Martingale CLT

We provide non-asymptotic convergence rates of the Polyak-Ruppert averag...
research
02/10/2021

Statistical Inference for Polyak-Ruppert Averaged Zeroth-order Stochastic Gradient Algorithm

As machine learning models are deployed in critical applications, it bec...

Please sign up or login with your details

Forgot password? Click here to reset