Q-Learning with Basic Emotions

09/06/2016
by   Wilfredo Badoy Jr., et al.
0

Q-learning is a simple and powerful tool in solving dynamic problems where environments are unknown. It uses a balance of exploration and exploitation to find an optimal solution to the problem. In this paper, we propose using four basic emotions: joy, sadness, fear, and anger to influence a Qlearning agent. Simulations show that the proposed affective agent requires lesser number of steps to find the optimal path. We found when affective agent finds the optimal path, the ratio between exploration to exploitation gradually decreases, indicating lower total step count in the long run

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/28/2019

Learning to Plan via Neural Exploration-Exploitation Trees

Sampling-based algorithms such as RRT and its variants are powerful tool...
research
02/09/2014

Recommandation mobile, sensible au contexte de contenus évolutifs: Contextuel-E-Greedy

We introduce in this paper an algorithm named Contextuel-E-Greedy that t...
research
12/05/2020

Exploration-Exploitation in Multi-Agent Learning: Catastrophe Theory Meets Game Theory

Exploration-exploitation is a powerful and practical tool in multi-agent...
research
09/13/2019

ISL: Optimal Policy Learning With Optimal Exploration-Exploitation Trade-Off

Traditionally, off-policy learning algorithms (such as Q-learning) and e...
research
10/30/2019

RBED: Reward Based Epsilon Decay

ε-greedy is a policy used to balance exploration and exploitation in man...
research
05/25/2014

HEPGAME and the Simplification of Expressions

Advances in high energy physics have created the need to increase comput...
research
08/16/2022

Solving the Diffusion of Responsibility Problem in Multiagent Reinforcement Learning with a Policy Resonance Approach

SOTA multiagent reinforcement algorithms distinguish themselves in many ...

Please sign up or login with your details

Forgot password? Click here to reset