Fast Thompson Sampling Algorithm with Cumulative Oversampling: Application to Budgeted Influence Maximization

04/24/2020
by   Shatian Wang, et al.
2

We propose a cumulative oversampling (CO) technique for Thompson Sampling (TS) that can be used to construct optimistic parameter estimates using significantly fewer samples from the posterior distributions compared to existing oversampling frameworks. We apply CO to a new budgeted variant of the Influence Maximization (IM) semi-bandits with linear generalization of edge weights. Combining CO with the oracle we designed for the offline problem, our online learning algorithm tackles the budget allocation, parameter learning, and reward maximization challenges simultaneously. We prove that our online learning algorithm achieves a scaled regret comparable to that of the UCB-based algorithms for IM semi-bandits. It is the first regret bound for TS-based algorithms for IM semi-bandits that does not depend linearly on the reciprocal of the minimum observation probability of an edge. In numerical experiments, our algorithm outperforms all UCB-based alternatives by a large margin.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2020

Thompson Sampling for Linearly Constrained Bandits

We address multi-armed bandits (MAB) where the objective is to maximize ...
research
02/16/2021

Optimal Algorithms for Private Online Learning in a Stochastic Environment

We consider two variants of private stochastic online learning. The firs...
research
05/21/2016

Online Influence Maximization under Independent Cascade Model with Semi-Bandit Feedback

We study the stochastic online problem of learning to influence in a soc...
research
06/09/2019

Factorization Bandits for Online Influence Maximization

We study the problem of online influence maximization in social networks...
research
06/17/2020

Analysis and Design of Thompson Sampling for Stochastic Partial Monitoring

We investigate finite stochastic partial monitoring, which is a general ...
research
11/12/2020

Online Influence Maximization under Linear Threshold Model

Online influence maximization (OIM) is a popular problem in social netwo...
research
09/06/2021

Online Learning of Independent Cascade Models with Node-level Feedback

We propose a detailed analysis of the online-learning problem for Indepe...

Please sign up or login with your details

Forgot password? Click here to reset