Log In Sign Up

Community Exploration: From Offline Optimization to Online Learning

by   Xiaowei Chen, et al.

We introduce the community exploration problem that has many real-world applications such as online advertising. In the problem, an explorer allocates limited budget to explore communities so as to maximize the number of members he could meet. We provide a systematic study of the community exploration problem, from offline optimization to online learning. For the offline setting where the sizes of communities are known, we prove that the greedy methods for both of non-adaptive exploration and adaptive exploration are optimal. For the online setting where the sizes of communities are not known and need to be learned from the multi-round explorations, we propose an `upper confidence' like algorithm that achieves the logarithmic regret bounds. By combining the feedback from different rounds, we can achieve a constant regret bound.


page 1

page 2

page 3

page 4


Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning

Multi-layered network exploration (MuLaNE) problem is an important probl...

Optimistic and Adaptive Lagrangian Hedging

In online learning an algorithm plays against an environment with losses...

Online Learning via Offline Greedy Algorithms: Applications in Market Design and Optimization

Motivated by online decision-making in time-varying combinatorial enviro...

BubbleRank: Safe Online Learning to Rerank

We study the problem of online learning to re-rank, where users provide ...

Delaytron: Efficient Learning of Multiclass Classifiers with Delayed Bandit Feedbacks

In this paper, we present online algorithm called Delaytron for learning...

History-Restricted Online Learning

We introduce the concept of history-restricted no-regret online learning...

Tiered Reinforcement Learning: Pessimism in the Face of Uncertainty and Constant Regret

We propose a new learning framework that captures the tiered structure o...