Minimax Strikes Back

12/19/2020
by   Quentin Cohen-Solal, et al.
0

Deep Reinforcement Learning (DRL) reaches a superhuman level of play in many complete information games. The state of the art search algorithm used in combination with DRL is Monte Carlo Tree Search (MCTS). We take another approach to DRL using a Minimax algorithm instead of MCTS and learning only the evaluation of states, not the policy. We show that for multiple games it is competitive with the state of the art DRL for the learning performances and for the confrontations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/30/2021

Deep Reinforcement Learning Aided Monte Carlo Tree Search for MIMO Detection

This paper proposes a novel multiple-input multiple-output (MIMO) symbol...
research
05/22/2020

Single-Agent Optimization Through Policy Iteration Using Monte-Carlo Tree Search

The combination of Monte-Carlo Tree Search (MCTS) and deep reinforcement...
research
02/01/2023

Alphazzle: Jigsaw Puzzle Solver with Deep Monte-Carlo Tree Search

Solving jigsaw puzzles requires to grasp the visual features of a sequen...
research
01/18/2017

First Study on Data Readiness Level

We introduce the idea of Data Readiness Level (DRL) to measure the relat...
research
09/24/2021

Combing Policy Evaluation and Policy Improvement in a Unified f-Divergence Framework

The framework of deep reinforcement learning (DRL) provides a powerful a...
research
10/10/2021

Vectorization of Raster Manga by Deep Reinforcement Learning

Manga is a popular Japanese-style comic form that consists of black-and-...
research
08/13/2019

Is Deep Reinforcement Learning Really Superhuman on Atari?

Consistent and reproducible evaluation of Deep Reinforcement Learning (D...

Please sign up or login with your details

Forgot password? Click here to reset