Multi-Agent Inverse Reinforcement Learning: Suboptimal Demonstrations and Alternative Solution Concepts

09/02/2021
by   Sage Bergerson, et al.
0

Multi-agent inverse reinforcement learning (MIRL) can be used to learn reward functions from agents in social environments. To model realistic social dynamics, MIRL methods must account for suboptimal human reasoning and behavior. Traditional formalisms of game theory provide computationally tractable behavioral models, but assume agents have unrealistic cognitive capabilities. This research identifies and compares mechanisms in MIRL methods which a) handle noise, biases and heuristics in agent decision making and b) model realistic equilibrium solution concepts. MIRL research is systematically reviewed to identify solutions for these challenges. The methods and results of these studies are analyzed and compared based on factors including performance accuracy, efficiency, and descriptive quality. We found that the primary methods for handling noise, biases and heuristics in MIRL were extensions of Maximum Entropy (MaxEnt) IRL to multi-agent settings. We also found that many successful solution concepts are generalizations of the traditional Nash Equilibrium (NE). These solutions include the correlated equilibrium, logistic stochastic best response equilibrium and entropy regularized mean field NE. Methods which use recursive reasoning or updating also perform well, including the feedback NE and archive multi-agent adversarial IRL. Success in modeling specific biases and heuristics in single-agent IRL and promising results using a Theory of Mind approach in MIRL imply that modeling specific biases and heuristics may be useful. Flexibility and unbiased inference in the identified alternative solution concepts suggest that a solution concept which has both recursive and generalized characteristics may perform well at modeling realistic social interactions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2021

Maximum Entropy Inverse Reinforcement Learning for Mean Field Games

Mean field games (MFG) facilitate the otherwise intractable reinforcemen...
research
01/26/2019

Multi-Agent Generalized Recursive Reasoning

We propose a new reasoning protocol called generalized recursive reasoni...
research
09/25/2020

Towards a Systematic Computational Framework for Modeling Multi-Agent Decision-Making at Micro Level for Smart Vehicles in a Smart World

We propose a multi-agent based computational framework for modeling deci...
research
01/26/2019

Probabilistic Recursive Reasoning for Multi-Agent Reinforcement Learning

Humans are capable of attributing latent mental contents such as beliefs...
research
03/25/2014

Multi-agent Inverse Reinforcement Learning for Zero-sum Games

In this paper we introduce a Bayesian framework for solving a class of p...
research
12/01/2022

Flexible social inference facilitates targeted social learning when rewards are not observable

Relying on others can be as risky as it can be rewarding. Advice seekers...

Please sign up or login with your details

Forgot password? Click here to reset