Finite-Time Analysis and Restarting Scheme for Linear Two-Time-Scale Stochastic Approximation

12/23/2019
by   Thinh T. Doan, et al.
0

Motivated by their broad applications in reinforcement learning, we study the linear two-time-scale stochastic approximation, an iterative method using two different step sizes for finding the solutions of a system of two equations. Our main focus is to characterize the finite-time complexity of this method under time-varying step sizes and Markovian noise. In particular, we show that the mean square errors of the variables generated by the method converge to zero at a sublinear rate O(k^2/3), where k is the number of iterations. We then improve the performance of this method by considering the restarting scheme, where we restart the algorithm after a predetermined number of iterations. We show that using this restarting method the complexity of the algorithm under time-varying step sizes is as good as the one using constant step sizes, but still achieving an exact converge to the desired solution. Moreover, the restarting scheme also helps to prevent the step sizes from getting too small, which is useful for the practical implementation of the linear two-time-scale stochastic approximation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/03/2020

Nonlinear Two-Time-Scale Stochastic Approximation: Convergence and Finite-Time Performance

Two-time-scale stochastic approximation, a generalized version of the po...
research
06/24/2020

Local Stochastic Approximation: A Unified View of Federated Learning and Distributed Multi-Task Reinforcement Learning Algorithms

Motivated by broad applications in reinforcement learning and federated ...
research
04/04/2021

Finite-Time Convergence Rates of Nonlinear Two-Time-Scale Stochastic Approximation under Markovian Noise

We study the so-called two-time-scale stochastic approximation, a simula...
research
09/29/2021

A Two-Time-Scale Stochastic Optimization Framework with Applications in Control and Reinforcement Learning

We study a novel two-time-scale stochastic gradient method for solving o...
research
11/24/2021

Finite-Time Error Bounds for Distributed Linear Stochastic Approximation

This paper considers a novel multi-agent linear stochastic approximation...
research
06/19/2022

Frank-Wolfe-based Algorithms for Approximating Tyler's M-estimator

Tyler's M-estimator is a well known procedure for robust and heavy-taile...
research
07/25/2022

Finite-Time Analysis of Asynchronous Q-learning under Diminishing Step-Size from Control-Theoretic View

Q-learning has long been one of the most popular reinforcement learning ...

Please sign up or login with your details

Forgot password? Click here to reset