DeepAI AI Chat
Log In Sign Up

Double Deep Q-Learning in Opponent Modeling

by   Yangtianze Tao, et al.
Tsinghua University

Multi-agent systems in which secondary agents with conflicting agendas also alter their methods need opponent modeling. In this study, we simulate the main agent's and secondary agents' tactics using Double Deep Q-Networks (DDQN) with a prioritized experience replay mechanism. Then, under the opponent modeling setup, a Mixture-of-Experts architecture is used to identify various opponent strategy patterns. Finally, we analyze our models in two environments with several agents. The findings indicate that the Mixture-of-Experts model, which is based on opponent modeling, performs better than DDQN.


page 1

page 2

page 3

page 4


Lenient Multi-Agent Deep Reinforcement Learning

A significant amount of research in recent years has been dedicated towa...

Variational Autoencoders for Opponent Modeling in Multi-Agent Systems

Multi-agent systems exhibit complex behaviors that emanate from the inte...

ViZDoom: DRQN with Prioritized Experience Replay, Double-Q Learning, & Snapshot Ensembling

ViZDoom is a robust, first-person shooter reinforcement learning environ...

Weighted Double Deep Multiagent Reinforcement Learning in Stochastic Cooperative Environments

Despite single agent deep reinforcement learning has achieved significan...

baller2vec++: A Look-Ahead Multi-Entity Transformer For Modeling Coordinated Agents

In many multi-agent spatiotemporal systems, the agents are under the inf...

Few self-involved agents among BC agents can lead to polarized local or global consensus

Social issues are generally discussed by highly-involved and less-involv...