Playing Flappy Bird via Asynchronous Advantage Actor Critic Algorithm

07/06/2019
by   Elit Cenk Alp, et al.
1

Flappy Bird, which has a very high popularity, has been trained in many algorithms. Some of these studies were trained from raw pixel values of game and some from specific attributes. In this study, the model was trained with raw game images, which had not been seen before. The trained model has learned as reinforcement when to make which decision. As an input to the model, the reward or penalty at the end of each step was returned and the training was completed. Flappy Bird game was trained with the Reinforcement Learning algorithm Deep Q-Network and Asynchronous Advantage Actor Critic (A3C) algorithms.

READ FULL TEXT

page 1

page 2

page 4

page 5

research
06/14/2021

Analysis of a Target-Based Actor-Critic Algorithm with Linear Function Approximation

Actor-critic methods integrating target networks have exhibited a stupen...
research
07/06/2018

End-to-End Race Driving with Deep Reinforcement Learning

We present research using the latest reinforcement learning algorithm fo...
research
04/13/2018

Robust Dual View Deep Agent

Motivated by recent advance of machine learning using Deep Reinforcement...
research
04/13/2018

Robust Dual View Depp Agent

Motivated by recent advance of machine learning using Deep Reinforcement...
research
02/12/2020

Fully Differentiable Procedural Content Generation through Generative Playing Networks

To procedurally create interactive content such as environments or game ...
research
04/08/2020

Solving the scalarization issues of Advantage-based Reinforcement Learning Algorithms

In this paper we investigate some of the issues that arise from the scal...
research
07/23/2018

Learning to Play Pong using Policy Gradient Learning

Activities in reinforcement learning (RL) revolve around learning the Ma...

Please sign up or login with your details

Forgot password? Click here to reset