DeepAI AI Chat
Log In Sign Up

Double Deep Q-Learning in Opponent Modeling

11/24/2022
by   Yangtianze Tao, et al.
Tsinghua University
0

Multi-agent systems in which secondary agents with conflicting agendas also alter their methods need opponent modeling. In this study, we simulate the main agent's and secondary agents' tactics using Double Deep Q-Networks (DDQN) with a prioritized experience replay mechanism. Then, under the opponent modeling setup, a Mixture-of-Experts architecture is used to identify various opponent strategy patterns. Finally, we analyze our models in two environments with several agents. The findings indicate that the Mixture-of-Experts model, which is based on opponent modeling, performs better than DDQN.

READ FULL TEXT

page 1

page 2

page 3

page 4

07/14/2017

Lenient Multi-Agent Deep Reinforcement Learning

A significant amount of research in recent years has been dedicated towa...
01/29/2020

Variational Autoencoders for Opponent Modeling in Multi-Agent Systems

Multi-agent systems exhibit complex behaviors that emanate from the inte...
01/03/2018

ViZDoom: DRQN with Prioritized Experience Replay, Double-Q Learning, & Snapshot Ensembling

ViZDoom is a robust, first-person shooter reinforcement learning environ...
02/23/2018

Weighted Double Deep Multiagent Reinforcement Learning in Stochastic Cooperative Environments

Despite single agent deep reinforcement learning has achieved significan...
04/24/2021

baller2vec++: A Look-Ahead Multi-Entity Transformer For Modeling Coordinated Agents

In many multi-agent spatiotemporal systems, the agents are under the inf...
05/25/2018

Few self-involved agents among BC agents can lead to polarized local or global consensus

Social issues are generally discussed by highly-involved and less-involv...