Modeling Social Group Communication with Multi-Agent Imitation Learning

03/04/2019
by   Navyata Sanghvi, et al.
10

In crowded social scenarios with a myriad of external stimuli, human brains exhibit a natural ability to filter out irrelevant information and narrowly focus their attention. In the midst of multiple groups of people, humans use such sensory gating to effectively further their own group's interactional goals. In this work, we consider the design of a policy network to model multi-group multi-person communication. Our policy takes as input the state of the world such as an agent's gaze direction, body pose of other agents or history of past actions, and outputs an optimal action such as speaking, listening or responding (communication modes). Inspired by humans' natural neurobiological filtering process, a central component of our policy network design is an information gating function, termed the Kinesic-Proxemic-Message Gate (KPM-Gate), that models the ability of an agent to selectively gather information from specific neighboring agents. The degree of influence of a neighbor is based on dynamic non-verbal cues such as body motion, head pose (kinesics) and interpersonal space (proxemics). We further show that the KPM-Gate can be used to discover social groups using its natural interpretation as a social attention mechanism. We pose the communication policy learning problem as a multi-agent imitation learning problem. We learn a single policy shared by all agents under the assumption of a decentralized Markov decision process. We term our policy network as the Multi-Agent Group Discovery and Communication Mode Network (MAGDAM network), as it learns social group structure in addition to the dynamics of group communication. Our experimental validation on both synthetic and real world data shows that our model is able to both discover social group structure and learn an accurate multi-agent communication policy.

READ FULL TEXT

page 1

page 2

page 3

page 5

page 7

page 10

research
07/10/2021

Multi-Agent Imitation Learning with Copulas

Multi-agent imitation learning aims to train multiple agents to perform ...
research
09/12/2021

Learning Selective Communication for Multi-Agent Path Finding

Learning communication via deep reinforcement learning (RL) or imitation...
research
02/28/2023

On the Role of Emergent Communication for Social Learning in Multi-Agent Reinforcement Learning

Explicit communication among humans is key to coordinating and learning....
research
07/07/2020

Policy learning with partial observation and mechanical constraints for multi-person modeling

Extracting the rules of real-world biological multi-agent behaviors is a...
research
11/05/2019

Learning to flock through reinforcement

Flocks of birds, schools of fish, insects swarms are examples of coordin...
research
02/23/2023

K-SHAP: Policy Clustering Algorithm for Anonymous State-Action Pairs

Learning agent behaviors from observational data has shown to improve ou...
research
06/26/2018

Learning Existing Social Conventions in Markov Games

In order for artificial agents to coordinate effectively with people, th...

Please sign up or login with your details

Forgot password? Click here to reset