An Asymptotically Optimal Strategy for Constrained Multi-armed Bandit Problems

05/03/2018
by   Hyeong Soo Chang, et al.
0

For the stochastic multi-armed bandit (MAB) problem from a constrained model that generalizes the classical one, we show that an asymptotic optimality is achievable by a simple strategy extended from the ϵ_t-greedy strategy. We provide a finite-time lower bound on the probability of correct selection of an optimal near-feasible arm that holds for all time steps. Under some conditions, the bound approaches one as time t goes to infinity. A particular example sequence of {ϵ_t} having the asymptotic convergence rate in the order of (1-1/t)^4 that holds from a sufficiently large t is also discussed.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset