D3C: Reducing the Price of Anarchy in Multi-Agent Learning

10/01/2020
by   Ian Gemp, et al.
0

Even in simple multi-agent systems, fixed incentives can lead to outcomes that are poor for the group and each individual agent. We propose a method, D3C, for online adjustment of agent incentives that reduces the loss incurred at a Nash equilibrium. Agents adjust their incentives by learning to mix their incentive with that of other agents, until a compromise is reached in a distributed fashion. We show that D3C improves outcomes for each agent and the group as a whole on several social dilemmas including a traffic network with Braess's paradox, a prisoner's dilemma, and several reinforcement learning domains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/11/2020

Multi-Agent Informational Learning Processes

We introduce a new mathematical model of multi-agent reinforcement learn...
research
04/06/2020

Using Multi-Agent Reinforcement Learning in Auction Simulations

Game theory has been developed by scientists as a theory of strategic in...
research
11/03/2022

Group Cohesion in Multi-Agent Scenarios as an Emergent Behavior

In this paper, we elaborate on the design and discuss the results of a m...
research
03/13/2018

Decentralised Learning in Systems with Many, Many Strategic Agents

Although multi-agent reinforcement learning can tackle systems of strate...
research
12/27/2022

Learning Individual Policies in Large Multi-agent Systems through Local Variance Minimization

In multi-agent systems with large number of agents, typically the contri...
research
09/26/1998

Learning Nested Agent Models in an Information Economy

We present our approach to the problem of how an agent, within an econom...
research
03/24/2021

Are energy savings the only reason for the emergence of bird echelon formation?

We analyze the conditions under which the emergence of frequently observ...

Please sign up or login with your details

Forgot password? Click here to reset