Implicit stochastic approximation

10/04/2015
by   Panos Toulis, et al.
0

The need to carry out parameter estimation from massive data has reinvigorated interest in iterative estimation methods, in statistics and machine learning. Classic work includes deterministic gradient-based methods, such as quasi-Newton, and stochastic gradient descent and its variants, including adaptive learning rates, acceleration and averaging. Current work increasingly relies on methods that employ proximal operators, leading to updates defined through implicit equations, which need to be solved at each iteration. Such methods are especially attractive in modern problems with massive data because they are numerically stable and converge with minimal assumptions, among other reasons. However, while the majority of existing methods can be subsumed into the gradient-free stochastic approximation framework developed by Robbins and Monro (1951), there is no such framework for methods with implicit updates. Here, we conceptualize a gradient-free implicit stochastic approximation procedure, and develop asymptotic and non-asymptotic theory for it. This new framework provides a theoretical foundation for gradient-based procedures that rely on implicit updates, and opens the door to iterative estimation methods that do not require a gradient, nor a fully known likelihood.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2015

Towards stability and optimality in stochastic gradient descent

Iterative procedures for parameter estimation based on stochastic gradie...
research
09/22/2015

Stochastic gradient descent methods for estimation with large data sets

We develop methods for parameter estimation in settings with large-scale...
research
06/14/2021

Smart Gradient – An Adaptive Technique for Improving Gradient Estimation

Computing the gradient of a function provides fundamental information ab...
research
03/19/2019

Convergence Analysis of Inexact Randomized Iterative Methods

In this paper we present a convergence rate analysis of inexact variants...
research
01/31/2023

Toward Efficient Gradient-Based Value Estimation

Gradient-based methods for value estimation in reinforcement learning ha...
research
03/06/2017

Forward and Reverse Gradient-Based Hyperparameter Optimization

We study two procedures (reverse-mode and forward-mode) for computing th...
research
07/11/2023

Implicit regularisation in stochastic gradient descent: from single-objective to two-player games

Recent years have seen many insights on deep learning optimisation being...

Please sign up or login with your details

Forgot password? Click here to reset