Interactive Restless Multi-armed Bandit Game and Swarm Intelligence Effect

03/13/2015
by   Shunsuke Yoshida, et al.
0

We obtain the conditions for the emergence of the swarm intelligence effect in an interactive game of restless multi-armed bandit (rMAB). A player competes with multiple agents. Each bandit has a payoff that changes with a probability p_c per round. The agents and player choose one of three options: (1) Exploit (a good bandit), (2) Innovate (asocial learning for a good bandit among n_I randomly chosen bandits), and (3) Observe (social learning for a good bandit). Each agent has two parameters (c,p_obs) to specify the decision: (i) c, the threshold value for Exploit, and (ii) p_obs, the probability for Observe in learning. The parameters (c,p_obs) are uniformly distributed. We determine the optimal strategies for the player using complete knowledge about the rMAB. We show whether or not social or asocial learning is more optimal in the (p_c,n_I) space and define the swarm intelligence effect. We conduct a laboratory experiment (67 subjects) and observe the swarm intelligence effect only if (p_c,n_I) are chosen so that social learning is far more optimal than asocial learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/17/2017

Online Multi-Armed Bandit

We introduce a novel variant of the multi-armed bandit problem, in which...
research
05/21/2019

Heterogeneous Stochastic Interactions for Multiple Agents in a Multi-armed Bandit Problem

We define and analyze a multi-agent multi-armed bandit problem in which ...
research
11/20/2018

Playing with and against Hedge

Hedge has been proposed as an adaptive scheme, which guides an agent's d...
research
07/23/2021

Finite-time Analysis of Globally Nonstationary Multi-Armed Bandits

We consider nonstationary multi-armed bandit problems where the model pa...
research
01/27/2023

Decentralized Online Bandit Optimization on Directed Graphs with Regret Bounds

We consider a decentralized multiplayer game, played over T rounds, with...
research
04/12/2018

Entangled photons for competitive multi-armed bandit problem: achievement of maximum social reward, equality, and deception prevention

The competitive multi-armed bandit (CMAB) problem is related to social i...
research
02/10/2021

Regression Oracles and Exploration Strategies for Short-Horizon Multi-Armed Bandits

This paper explores multi-armed bandit (MAB) strategies in very short ho...

Please sign up or login with your details

Forgot password? Click here to reset