First-order regret bounds for combinatorial semi-bandits

02/23/2015
by   Gergely Neu, et al.
0

We consider the problem of online combinatorial optimization under semi-bandit feedback, where a learner has to repeatedly pick actions from a combinatorial decision set in order to minimize the total losses associated with its decisions. After making each decision, the learner observes the losses associated with its action, but not other losses. For this problem, there are several learning algorithms that guarantee that the learner's expected regret grows as O(√(T)) with the number of rounds T. In this paper, we propose an algorithm that improves this scaling to O(√(L_T^*)), where L_T^* is the total loss of the best action. Our algorithm is among the first to achieve such guarantees in a partial-feedback scheme, and the first one to do so in a combinatorial setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/30/2014

Nonstochastic Multi-Armed Bandits with Graph-Structured Feedback

We present and study a partial-information model of online learning, whe...
research
11/09/2017

Small-loss bounds for online learning with partial information

We consider the problem of adversarial (non-stochastic) online learning ...
research
10/05/2020

An Efficient Algorithm for Cooperative Semi-Bandits

We consider the problem of asynchronous online combinatorial optimizatio...
research
03/17/2015

Importance weighting without importance weights: An efficient algorithm for combinatorial semi-bandits

We propose a sample-efficient alternative for importance weighting for s...
research
02/08/2019

Bandit Principal Component Analysis

We consider a partial-feedback variant of the well-studied online PCA pr...
research
04/23/2022

Smoothed Online Combinatorial Optimization Using Imperfect Predictions

Smoothed online combinatorial optimization considers a learner who repea...
research
06/19/2022

Nested bandits

In many online decision processes, the optimizing agent is called to cho...

Please sign up or login with your details

Forgot password? Click here to reset