On the Linear Convergence of Policy Gradient under Hadamard Parameterization

05/31/2023
by   Jiacai Liu, et al.
0

The convergence of deterministic policy gradient under the Hadamard parametrization is studied in the tabular setting and the global linear convergence of the algorithm is established. To this end, we first show that the error decreases at an O(1/k) rate for all the iterations. Based on this result, we further show that the algorithm has a faster local linear convergence rate after k_0 iterations, where k_0 is a constant that only depends on the MDP problem and the step size. Overall, the algorithm displays a linear convergence rate for all the iterations with a loose constant than that for the local linear convergence rate.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/30/2022

Linear Convergence for Natural Policy Gradient with Log-linear Policy Parametrization

We analyze the convergence rate of the unregularized natural policy grad...
research
05/24/2023

Adaptive Policy Learning to Additional Tasks

This paper develops a policy learning method for tuning a pre-trained po...
research
03/20/2018

Frank-Wolfe with Subsampling Oracle

We analyze two novel randomized variants of the Frank-Wolfe (FW) or cond...
research
04/02/2021

Neurons learn slower than they think

Recent studies revealed complex convergence dynamics in gradient-based m...
research
04/16/2021

Distributed TD(0) with Almost No Communication

We provide a new non-asymptotic analysis of distributed TD(0) with linea...
research
08/31/2020

Super-linear convergence in the p-adic QR-algorithm

The QR-algorithm is one of the most important algorithms in linear algeb...
research
04/01/2017

Faster Subgradient Methods for Functions with Hölderian Growth

The purpose of this manuscript is to derive new convergence results for ...

Please sign up or login with your details

Forgot password? Click here to reset