Player Modeling via Multi-Armed Bandits

02/10/2021
by   Robert C. Gray, et al.
0

This paper focuses on building personalized player models solely from player behavior in the context of adaptive games. We present two main contributions: The first is a novel approach to player modeling based on multi-armed bandits (MABs). This approach addresses, at the same time and in a principled way, both the problem of collecting data to model the characteristics of interest for the current player and the problem of adapting the interactive experience based on this model. Second, we present an approach to evaluating and fine-tuning these algorithms prior to generating data in a user study. This is an important problem, because conducting user studies is an expensive and labor-intensive process; therefore, an ability to evaluate the algorithms beforehand can save a significant amount of resources. We evaluate our approach in the context of modeling players' social comparison orientation (SCO) and present empirical results from both simulations and real players.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/28/2019

Non-Stochastic Multi-Player Multi-Armed Bandits: Optimal Rate With Collision Information, Sublinear Without

We consider the non-stochastic version of the (cooperative) multi-player...
research
12/12/2022

Decentralized Stochastic Multi-Player Multi-Armed Walking Bandits

Multi-player multi-armed bandit is an increasingly relevant decision-mak...
research
11/08/2021

An Instance-Dependent Analysis for the Cooperative Multi-Player Multi-Armed Bandit

We study the problem of information sharing and cooperation in Multi-Pla...
research
02/18/2023

Improving Fairness in Adaptive Social Exergames via Shapley Bandits

Algorithmic fairness is an essential requirement as AI becomes integrate...
research
05/30/2023

Competing for Shareable Arms in Multi-Player Multi-Armed Bandits

Competitions for shareable and limited resources have long been studied ...
research
11/02/2020

On No-Sensing Adversarial Multi-player Multi-armed Bandits with Collision Communications

We study the notoriously difficult no-sensing adversarial multi-player m...
research
05/15/2020

SEAM methodology for context-rich player matchup evaluations

We develop the SEAM (synthetic estimated average matchup) method for des...

Please sign up or login with your details

Forgot password? Click here to reset