Semi-bandit Optimization in the Dispersed Setting

04/18/2019
by   Maria-Florina Balcan, et al.
0

In this work, we study the problem of online optimization of piecewise Lipschitz functions with semi-bandit feedback. This challenging class of non-convex optimization problems often arises in algorithm selection problems for combinatorial settings, where the goal is to find the best algorithm from a large algorithm family for a specific application domain. In these settings, each evaluation of the loss functions in the optimization problem can be computationally expensive, often requiring the learner to run a combinatorial algorithm to measure its performance. Combined with the fact that small differences between similar algorithms in the family can lead to cascading changes in algorithm behavior, efficient online optimization in these settings is a challenging problem. However, we show that in many applications, evaluating the loss function for one algorithm choice can sometimes reveal the loss for a range of similar algorithms, essentially for free. We develop online optimization algorithms capable of using this kind of extra information by working in the semi-bandit feedback setting. Our algorithms achieve regret bounds that are essentially as good as algorithms under full-information feedback and are significantly more computationally efficient. We apply our semi-bandit optimization results to obtain online algorithm selection procedures for two rich families of combinatorial algorithms. We provide the first provable guarantees for online algorithm selection for clustering problems using a family of clustering algorithms containing classic linkage procedures. We also show how to select algorithms from a family of greedy knapsack algorithms with simultaneously lower computational complexity and stronger regret bounds than the best algorithm selection procedures from prior work.

READ FULL TEXT
research
04/20/2012

Regret in Online Combinatorial Optimization

We address online linear optimization problems when the possible actions...
research
02/25/2017

Efficient Online Bandit Multiclass Learning with Õ(√(T)) Regret

We present an efficient second-order algorithm with Õ(1/η√(T)) regret fo...
research
07/22/2019

Online optimization of piecewise Lipschitz functions in changing environments

In an online optimization problem we are required to choose a sequence o...
research
06/01/2019

Adaptive Online Learning for Gradient-Based Optimizers

As application demands for online convex optimization accelerate, the ne...
research
04/18/2018

Online Non-Additive Path Learning under Full and Partial Information

We consider the online path learning problem in a graph with non-additiv...
research
02/12/2022

Adaptive Bandit Convex Optimization with Heterogeneous Curvature

We consider the problem of adversarial bandit convex optimization, that ...
research
01/08/2023

Fully Dynamic Online Selection through Online Contention Resolution Schemes

We study fully dynamic online selection problems in an adversarial/stoch...

Please sign up or login with your details

Forgot password? Click here to reset