Statistical Inference for the Population Landscape via Moment Adjusted Stochastic Gradients

12/20/2017
by   Tengyuan Liang, et al.
0

Modern statistical inference tasks often require iterative optimization methods to approximate the solution. Convergence analysis from optimization only tells us how well we are approximating the solution deterministically, but overlooks the sampling nature of the data. However, due to the randomness in the data, statisticians are keen to provide uncertainty quantification, or confidence, for the answer obtained after certain steps of optimization. Therefore, it is important yet challenging to understand the sampling distribution of the iterative optimization methods. This paper makes some progress along this direction by introducing a new stochastic optimization method for statistical inference, the moment adjusted stochastic gradient descent. We establish non-asymptotic theory that characterizes the statistical distribution of the iterative methods, with good optimization guarantee. On the statistical front, the theory allows for model misspecification, with very mild conditions on the data. For optimization, the theory is flexible for both the convex and non-convex cases. Remarkably, the moment adjusting idea motivated from "error standardization" in statistics achieves similar effect as Nesterov's acceleration in optimization, for certain convex problems as in fitting generalized linear models. We also demonstrate this acceleration effect in the non-convex setting through experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/05/2021

Online Statistical Inference for Gradient-free Stochastic Optimization

As gradient-free stochastic optimization gains emerging attention for a ...
research
03/21/2022

A Local Convergence Theory for the Stochastic Gradient Descent Method in Non-Convex Optimization With Non-isolated Local Minima

Non-convex loss functions arise frequently in modern machine learning, a...
research
11/06/2018

Double Adaptive Stochastic Gradient Optimization

Adaptive moment methods have been remarkably successful in deep learning...
research
01/30/2022

SRKCD: a stabilized Runge-Kutta method for stochastic optimization

We introduce a family of stochastic optimization methods based on the Ru...
research
01/10/2022

Non-Asymptotic Guarantees for Robust Statistical Learning under (1+ε)-th Moment Assumption

There has been a surge of interest in developing robust estimators for m...
research
02/19/2018

On the Optimization of Deep Networks: Implicit Acceleration by Overparameterization

Conventional wisdom in deep learning states that increasing depth improv...
research
06/12/2018

INFERNO: Inference-Aware Neural Optimisation

Complex computer simulations are commonly required for accurate data mod...

Please sign up or login with your details

Forgot password? Click here to reset