Online Linear Optimization with Many Hints

10/06/2020
by   Aditya Bhaskara, et al.
0

We study an online linear optimization (OLO) problem in which the learner is provided access to K "hint" vectors in each round prior to making a decision. In this setting, we devise an algorithm that obtains logarithmic regret whenever there exists a convex combination of the K hints that has positive correlation with the cost vectors. This significantly extends prior work that considered only the case K=1. To accomplish this, we develop a way to combine many arbitrary OLO algorithms to obtain regret only a logarithmically worse factor than the minimum regret of the original algorithms in hindsight; this result is of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2019

Matrix-Free Preconditioning in Online Learning

We provide an online convex optimization algorithm with regret that inte...
research
02/11/2020

Online Learning with Imperfect Hints

We consider a variant of the classical online linear optimization proble...
research
02/15/2022

Damped Online Newton Step for Portfolio Selection

We revisit the classic online portfolio selection problem, where at each...
research
01/30/2023

Improved Regret for Efficient Online Reinforcement Learning with Linear Function Approximation

We study reinforcement learning with linear function approximation and a...
research
02/27/2020

Lipschitz and Comparator-Norm Adaptivity in Online Learning

We study Online Convex Optimization in the unbounded setting where neith...
research
02/20/2019

Adaptive scale-invariant online algorithms for learning linear models

We consider online learning with linear models, where the algorithm pred...
research
04/03/2020

Universal Algorithms: Beyond the Simplex

The bulk of universal algorithms in the online convex optimisation liter...

Please sign up or login with your details

Forgot password? Click here to reset