A Deep Q-Learning Agent for the L-Game with Variable Batch Training

02/17/2018
by   Petros Giannakopoulos, et al.
0

We employ the Deep Q-Learning algorithm with Experience Replay to train an agent capable of achieving a high-level of play in the L-Game while self-learning from low-dimensional states. We also employ variable batch size for training in order to mitigate the loss of the rare reward signal and significantly accelerate training. Despite the large action space due to the number of possible moves, the low-dimensional state space and the rarity of rewards, which only come at the end of a game, DQL is successful in training an agent capable of strong play without the use of any search methods or domain knowledge.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2022

Supervised and Reinforcement Learning from Observations in Reconnaissance Blind Chess

In this work, we adapt a training approach inspired by the original Alph...
research
05/18/2018

Solving the Rubik's Cube Without Human Knowledge

A generally intelligent agent must be able to teach itself how to solve ...
research
10/21/2021

Locality-Sensitive Experience Replay for Online Recommendation

Online recommendation requires handling rapidly changing user preference...
research
05/06/2023

A Novel Reward Shaping Function for Single-Player Mahjong

Mahjong is a complex game with an intractably large state space with ext...
research
09/10/2020

Using Graph Convolutional Networks and TD(λ) to play the game of Risk

Risk is 6 player game with significant randomness and a large game-tree ...
research
11/07/2022

Are AlphaZero-like Agents Robust to Adversarial Perturbations?

The success of AlphaZero (AZ) has demonstrated that neural-network-based...
research
06/14/2020

Tackling Morpion Solitaire with AlphaZero-likeRanked Reward Reinforcement Learning

Morpion Solitaire is a popular single player game, performed with paper ...

Please sign up or login with your details

Forgot password? Click here to reset