PAC Reinforcement Learning Algorithm for General-Sum Markov Games

09/05/2020
by   Ashkan Zehfroosh, et al.
0

This paper presents a theoretical framework for probably approximately correct (PAC) multi-agent reinforcement learning (MARL) algorithms for Markov games. The paper offers an extension to the well-known Nash Q-learning algorithm, using the idea of delayed Q-learning, in order to build a new PAC MARL algorithm for general-sum Markov games. In addition to guiding the design of a provably PAC MARL algorithm, the framework enables checking whether an arbitrary MARL algorithm is PAC. Comparative numerical results demonstrate performance and robustness.

READ FULL TEXT

page 9

page 10

research
09/05/2020

A Hybrid PAC Reinforcement Learning Algorithm

This paper offers a new hybrid probably asymptotically correct (PAC) rei...
research
04/17/2023

Scenario Approach for Parametric Markov Models

In this paper, we propose an approximating framework for analyzing param...
research
03/22/2017

Unifying PAC and Regret: Uniform PAC Bounds for Episodic Reinforcement Learning

Statistical performance bounds for reinforcement learning (RL) algorithm...
research
05/10/2019

PAC Statistical Model Checking for Markov Decision Processes and Stochastic Games

Statistical model checking (SMC) is a technique for analysis of probabil...
research
11/24/2021

Reinforcement Learning for General LTL Objectives Is Intractable

In recent years, researchers have made significant progress in devising ...
research
08/13/2018

On Passivity, Reinforcement Learning and Higher-Order Learning in Multi-Agent Finite Games

In this paper, we propose a passivity-based methodology for analysis and...
research
03/20/2019

A Learning Framework for Distribution-Based Game-Theoretic Solution Concepts

The past few years have seen several works establishing PAC frameworks f...

Please sign up or login with your details

Forgot password? Click here to reset