Recursive Experts: An Efficient Optimal Mixture of Learning Systems in Dynamic Environments

09/19/2020
by   Kaan Gokcesu, et al.
0

Sequential learning systems are used in a wide variety of problems from decision making to optimization, where they provide a 'belief' (opinion) to nature, and then update this belief based on the feedback (result) to minimize (or maximize) some cost or loss (conversely, utility or gain). The goal is to reach an objective by exploiting the temporal relation inherent to the nature's feedback (state). By exploiting this relation, specific learning systems can be designed that perform asymptotically optimal for various applications. However, if the framework of the problem is not stationary, i.e., the nature's state sometimes changes arbitrarily, the past cumulative belief revision done by the system may become useless and the system may fail if it lacks adaptivity. While this adaptivity can be directly implemented in specific cases (e.g., convex optimization), it is mostly not straightforward for general learning tasks. To this end, we propose an efficient optimal mixture framework for general sequential learning systems, which we call the recursive experts for dynamic environments. For this purpose, we design hyper-experts that incorporate the learning systems at our disposal and recursively merge in a specific way to achieve minimax optimal regret bounds up to constant factors. The multiplicative increases in computational complexity from the initial system to our adaptive system are only logarithmic-in-time factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/29/2019

Bandit Convex Optimization in Non-stationary Environments

Bandit Convex Optimization (BCO) is a fundamental framework for modeling...
research
07/07/2020

Dynamic Regret of Convex and Smooth Functions

We investigate online convex optimization in non-stationary environments...
research
02/20/2020

Optimal anytime regret with two experts

The multiplicative weights method is an algorithm for the problem of pre...
research
02/15/2020

Jelly Bean World: A Testbed for Never-Ending Learning

Machine learning has shown growing success in recent years. However, cur...
research
12/16/2022

An Efficient Framework for Monitoring Subgroup Performance of Machine Learning Systems

Monitoring machine learning systems post deployment is critical to ensur...
research
08/13/2021

Optimal and Efficient Algorithms for General Mixable Losses against Switching Oracles

We investigate the problem of online learning, which has gained signific...
research
04/02/2014

From ADP to the Brain: Foundations, Roadmap, Challenges and Research Priorities

This paper defines and discusses Mouse Level Computational Intelligence ...

Please sign up or login with your details

Forgot password? Click here to reset