Bandit Models of Human Behavior: Reward Processing in Mental Disorders

06/07/2017
by   Djallel Bouneffouf, et al.
0

Drawing an inspiration from behavioral studies of human decision making, we propose here a general parametric framework for multi-armed bandit problem, which extends the standard Thompson Sampling approach to incorporate reward processing biases associated with several neurological and psychiatric conditions, including Parkinson's and Alzheimer's diseases, attention-deficit/hyperactivity disorder (ADHD), addiction, and chronic pain. We demonstrate empirically that the proposed parametric approach can often outperform the baseline Thompson Sampling on a variety of datasets. Moreover, from the behavioral modeling perspective, our parametric framework can be viewed as a first step towards a unifying computational model capturing reward processing abnormalities across multiple mental conditions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/21/2019

Reinforcement Learning Models of Human Behavior: Reward Processing in Mental Disorders

Drawing an inspiration from behavioral studies of human decision making,...
research
06/21/2019

Split Q Learning: Reinforcement Learning with Two-Stream Rewards

Drawing an inspiration from behavioral studies of human decision making,...
research
05/10/2020

Unified Models of Human Behavioral Agents in Bandits, Contextual Bandits and RL

Artificial behavioral agents are often evaluated based on their consiste...
research
01/20/2023

GBOSE: Generalized Bandit Orthogonalized Semiparametric Estimation

In sequential decision-making scenarios i.e., mobile health recommendati...
research
04/12/2021

Risk-Averse Biased Human Policies in Assistive Multi-Armed Bandit Settings

Assistive multi-armed bandit problems can be used to model team situatio...
research
02/21/2023

Learning signatures of decision making from many individuals playing the same game

Human behavior is incredibly complex and the factors that drive decision...
research
05/19/2022

Multi-Armed Bandits in Brain-Computer Interfaces

The multi-armed bandit (MAB) problem models a decision-maker that optimi...

Please sign up or login with your details

Forgot password? Click here to reset