Model Selection in Contextual Stochastic Bandit Problems

03/03/2020 ∙ by Aldo Pacchiano, et al. ∙ 0

We study model selection in stochastic bandit problems. Our approach relies on a master algorithm that selects its actions among candidate base algorithms. While this problem is studied for specific classes of stochastic base algorithms, our objective is to provide a method that can work with more general classes of stochastic base algorithms. We propose a master algorithm inspired by CORRAL <cit.> and introduce a novel and generic smoothing transformation for stochastic bandit algorithms that permits us to obtain O(√(T)) regret guarantees for a wide class of base algorithms when working along with our master. We exhibit a lower bound showing that even when one of the base algorithms has O(log T) regret, in general it is impossible to get better than Ω(√(T)) regret in model selection, even asymptotically. We apply our algorithm to choose among different values of ϵ for the ϵ-greedy algorithm, and to choose between the k-armed UCB and linear UCB algorithms. Our empirical studies further confirm the effectiveness of our model-selection method.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.