Opponent Aware Reinforcement Learning

08/22/2019
by   Víctor Gallego, et al.
5

In several reinforcement learning (RL) scenarios such as security settings, there may be adversaries trying to interfere with the reward generating process for their own benefit. We introduce Threatened Markov Decision Processes (TMDPs) as a framework to support an agent against potential opponents in a RL context. We also propose a level-k thinking scheme resulting in a novel learning approach to deal with TMDPs. After introducing our framework and deriving theoretical results, relevant empirical evidence is given via extensive experiments, showing the benefits of accounting for adversaries in RL while the agent learns

READ FULL TEXT

page 14

page 21

research
09/05/2018

Reinforcement Learning under Threats

In several reinforcement learning (RL) scenarios, mainly in security set...
research
01/15/2020

Lipschitz Lifelong Reinforcement Learning

We consider the problem of knowledge transfer when an agent is facing a ...
research
05/15/2001

Market-Based Reinforcement Learning in Partially Observable Worlds

Unlike traditional reinforcement learning (RL), market-based RL is in pr...
research
06/20/2023

Reward Shaping via Diffusion Process in Reinforcement Learning

Reinforcement Learning (RL) models have continually evolved to navigate ...
research
10/25/2019

On the convergence of projective-simulation-based reinforcement learning in Markov decision processes

In recent years, the interest in leveraging quantum effects for enhancin...
research
10/07/2022

Knowledge-Grounded Reinforcement Learning

Receiving knowledge, abiding by laws, and being aware of regulations are...
research
10/03/2022

Mastering Spatial Graph Prediction of Road Networks

Accurately predicting road networks from satellite images requires a glo...

Please sign up or login with your details

Forgot password? Click here to reset