Learning through Probing: a decentralized reinforcement learning architecture for social dilemmas

09/26/2018
by   Nicolas Anastassacos, et al.
0

Multi-agent reinforcement learning has received significant interest in recent years notably due to the advancements made in deep reinforcement learning which have allowed for the developments of new architectures and learning algorithms. Using social dilemmas as the training ground, we present a novel learning architecture, Learning through Probing (LTP), where agents utilize a probing mechanism to incorporate how their opponent's behavior changes when an agent takes an action. We use distinct training phases and adjust rewards according to the overall outcome of the experiences accounting for changes to the opponents behavior. We introduce a parameter eta to determine the significance of these future changes to opponent behavior. When applied to the Iterated Prisoner's Dilemma (IPD), LTP agents demonstrate that they can learn to cooperate with each other, achieving higher average cumulative rewards than other reinforcement learning methods while also maintaining good performance in playing against static agents that are present in Axelrod tournaments. We compare this method with traditional reinforcement learning algorithms and agent-tracking techniques to highlight key differences and potential applications. We also draw attention to the differences between solving games and societal-like interactions and analyze the training of Q-learning agents in makeshift societies. This is to emphasize how cooperation may emerge in societies and demonstrate this using environments where interactions with opponents are determined through a random encounter format of the IPD.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/18/2022

Credit-cognisant reinforcement learning for multi-agent cooperation

Traditional multi-agent reinforcement learning (MARL) algorithms, such a...
research
05/21/2022

Co-design of Embodied Neural Intelligence via Constrained Evolution

We introduce a novel co-design method for autonomous moving agents' shap...
research
06/09/2021

Deception in Social Learning: A Multi-Agent Reinforcement Learning Perspective

Within the framework of Multi-Agent Reinforcement Learning, Social Learn...
research
10/19/2018

Social Influence as Intrinsic Motivation for Multi-Agent Deep Reinforcement Learning

We propose a unified mechanism for achieving coordination and communicat...
research
07/07/2018

How game complexity affects the playing behavior of synthetic agents

Agent based simulation of social organizations, via the investigation of...
research
11/09/2020

Testbeds for Reinforcement Learning

We present three problems modeled after animal learning experiments desi...
research
05/10/2023

Learning Optimal "Pigovian Tax" in Sequential Social Dilemmas

In multi-agent reinforcement learning, each agent acts to maximize its i...

Please sign up or login with your details

Forgot password? Click here to reset