Multi-Agent Asynchronous Cooperation with Hierarchical Reinforcement Learning

03/29/2022
by   Xubo Lyu, et al.
1

Hierarchical multi-agent reinforcement learning (MARL) has shown a significant learning efficiency by searching policy over higher-level, temporally extended actions (options). However, standard policy gradient-based MARL methods have a difficulty generalizing to option-based scenarios due to the asynchronous executions of multi-agent options. In this work, we propose a mathematical framework to enable policy gradient optimization over asynchronous multi-agent options by adjusting option-based policy distribution as well as trajectory probability. We study our method under a set of multi-agent cooperative setups with varying inter-dependency levels, and evaluate the effectiveness of our method on typical option-based multi-agent cooperation tasks.

READ FULL TEXT

page 10

page 17

page 18

research
10/21/2019

Multi-agent Hierarchical Reinforcement Learning with Dynamic Termination

In a multi-agent system, an agent's optimal policy will typically depend...
research
06/04/2019

Options as responses: Grounding behavioural hierarchies in multi-agent RL

We propose a novel hierarchical agent architecture for multi-agent reinf...
research
07/05/2023

Multi-Agent Cooperation via Unsupervised Learning of Joint Intentions

The field of cooperative multi-agent reinforcement learning (MARL) has s...
research
09/13/2023

Characterizing Speed Performance of Multi-Agent Reinforcement Learning

Multi-Agent Reinforcement Learning (MARL) has achieved significant succe...
research
11/20/2017

Situationally Aware Options

Hierarchical abstractions, also known as options -- a type of temporally...
research
01/20/2022

Multi-agent Covering Option Discovery based on Kronecker Product of Factor Graphs

Covering option discovery has been developed to improve the exploration ...

Please sign up or login with your details

Forgot password? Click here to reset