Convergence of Recursive Stochastic Algorithms using Wasserstein Divergence

03/25/2020
by   Abhishek Gupta, et al.
9

This paper develops a unified framework, based on iterated random operator theory, to analyze the convergence of constant stepsize recursive stochastic algorithms (RSAs) in machine learning and reinforcement learning. RSAs use randomization to efficiently compute expectations, and so their iterates form a stochastic process. The key idea is to lift the RSA into an appropriate higher-dimensional space and then express it as an equivalent Markov chain. Instead of determining the convergence of this Markov chain (which may not converge under constant stepsize), we study the convergence of the distribution of this Markov chain. To study this, we define a new notion of Wasserstein divergence. We show that if the distribution of the iterates in the Markov chain satisfy certain contraction property with respect to the Wasserstein divergence, then the Markov chain admits an invariant distribution. Inspired by the SVRG algorithm, we develop a method to convert any RSA to a variance reduced RSA that converges to the optimal solution with in almost sure sense or in probability. We show that convergence of a large family of constant stepsize RSAs can be understood using this framework. We apply this framework to ascertain the convergence of mini-batch SGD, forward-backward splitting with catalyst, SVRG, SAGA, empirical Q value iteration, synchronous Q-learning, enhanced policy iteration, and MDPs with a generative model. We also develop two new algorithms for reinforcement learning and establish their convergence using this framework.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/17/2016

Convergence of Contrastive Divergence Algorithm in Exponential Family

This paper studies the convergence properties of contrastive divergence ...
research
10/03/2022

Bias and Extrapolation in Markovian Linear Stochastic Approximation with Constant Stepsizes

We consider Linear Stochastic Approximation (LSA) with a constant stepsi...
research
04/24/2019

Some Limit Properties of Markov Chains Induced by Stochastic Recursive Algorithms

Recursive stochastic algorithms have gained significant attention in the...
research
04/21/2018

Global Convergence Analysis of the Flower Pollination Algorithm: A Discrete-Time Markov Chain Approach

Flower pollination algorithm is a recent metaheuristic algorithm for sol...
research
11/19/2021

Learn Quasi-stationary Distributions of Finite State Markov Chain

We propose a reinforcement learning (RL) approach to compute the express...
research
10/06/2020

Reinforcement Learning in Deep Structured Teams: Initial Results with Finite and Infinite Valued Features

In this paper, we consider Markov chain and linear quadratic models for ...
research
12/06/2020

Contrastive Divergence Learning is a Time Reversal Adversarial Game

Contrastive divergence (CD) learning is a classical method for fitting u...

Please sign up or login with your details

Forgot password? Click here to reset