Concentration of Contractive Stochastic Approximation and Reinforcement Learning

06/27/2021
by   Siddharth Chandak, et al.
0

Using a martingale concentration inequality, concentration bounds `from time n_0 on' are derived for stochastic approximation algorithms with contractive maps and both martingale difference and Markov noises. These are applied to reinforcement learning algorithms, in particular to asynchronous Q-learning and TD(0).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/03/2023

A Tutorial Introduction to Reinforcement Learning

In this paper, we present a brief survey of Reinforcement Learning (RL),...
research
12/08/2020

A Concentration Inequality for the Facility Location Problem

We give a concentration inequality for a stochastic version of the facil...
research
03/28/2023

Concentration of Contractive Stochastic Approximation: Additive and Multiplicative Noise

In this work, we study the concentration behavior of a stochastic approx...
research
09/08/2021

Convergence of Batch Asynchronous Stochastic Approximation With Applications to Reinforcement Learning

The stochastic approximation (SA) algorithm is a widely used probabilist...
research
09/18/2020

Deviation bound for non-causal machine learning

Concentration inequality are widely used for analysing machines learning...
research
05/19/2017

Atari games and Intel processors

The asynchronous nature of the state-of-the-art reinforcement learning a...
research
10/09/2022

A Concentration Bound for Distributed Stochastic Approximation

We revisit the classical model of Tsitsiklis, Bertsekas and Athans for d...

Please sign up or login with your details

Forgot password? Click here to reset