Near Optimal Memory-Regret Tradeoff for Online Learning

03/03/2023
by   Binghui Peng, et al.
0

In the experts problem, on each of T days, an agent needs to follow the advice of one of n “experts”. After each day, the loss associated with each expert's advice is revealed. A fundamental result in learning theory says that the agent can achieve vanishing regret, i.e. their cumulative loss is within o(T) of the cumulative loss of the best-in-hindsight expert. Can the agent perform well without sufficient space to remember all the experts? We extend a nascent line of research on this question in two directions: ∙ We give a new algorithm against the oblivious adversary, improving over the memory-regret tradeoff obtained by [PZ23], and nearly matching the lower bound of [SWXZ22]. ∙ We also consider an adaptive adversary who can observe past experts chosen by the agent. In this setting we give both a new algorithm and a novel lower bound, proving that roughly √(n) memory is both necessary and sufficient for obtaining o(T) regret.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/02/2023

Improved Space Bounds for Learning with Experts

We give improved tradeoffs between space and regret for the online learn...
research
07/16/2022

Online Prediction in Sub-linear Space

We provide the first sub-linear space and sub-linear regret algorithm fo...
research
03/03/2023

Streaming Algorithms for Learning with Experts: Deterministic Versus Robust

In the online learning with experts problem, an algorithm must make a pr...
research
04/21/2022

Memory Bounds for the Experts Problem

Online learning with expert advice is a fundamental problem of sequentia...
research
01/11/2022

Learning what to remember

We consider a lifelong learning scenario in which a learner faces a neve...
research
10/27/2020

Online Learning with Primary and Secondary Losses

We study the problem of online learning with primary and secondary losse...
research
11/14/2012

Distributed Non-Stochastic Experts

We consider the online distributed non-stochastic experts problem, where...

Please sign up or login with your details

Forgot password? Click here to reset