Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex Cost Functions

02/25/2022
by   Nima Eshraghi, et al.
0

The performance of online convex optimization algorithms in a dynamic environment is often expressed in terms of the dynamic regret, which measures the decision maker's performance against a sequence of time-varying comparators. In the analysis of the dynamic regret, prior works often assume Lipschitz continuity or uniform smoothness of the cost functions. However, there are many important cost functions in practice that do not satisfy these conditions. In such cases, prior analyses are not applicable and fail to guarantee the optimization performance. In this letter, we show that it is possible to bound the dynamic regret, even when neither Lipschitz continuity nor uniform smoothness is present. We adopt the notion of relative smoothness with respect to some user-defined regularization function, which is a much milder requirement on the cost functions. We first show that under relative smoothness, the dynamic regret has an upper bound based on the path length and functional variation. We then show that with an additional condition of relatively strong convexity, the dynamic regret can be bounded by the path length and gradient variation. These regret bounds provide performance guarantees to a wide variety of online optimization problems that arise in different application domains. Finally, we present numerical experiments that demonstrate the advantage of adopting a regularization function under which the cost functions are relatively smooth.

READ FULL TEXT
research
10/22/2020

Regret Bounds without Lipschitz Continuity: Online Learning with Relative-Lipschitz Losses

In online convex optimization (OCO), Lipschitz continuity of the functio...
research
06/30/2019

Efficient Online Convex Optimization with Adaptively Minimax Optimal Dynamic Regret

We introduce an online convex optimization algorithm using projected sub...
research
02/11/2023

Improved Dynamic Regret for Online Frank-Wolfe

To deal with non-stationary online problems with complex constraints, we...
research
07/06/2022

Online Bilevel Optimization: Regret Analysis of Online Alternating Gradient Methods

Online optimization is a well-established optimization paradigm that aim...
research
03/28/2022

Optimistic Online Convex Optimization in Dynamic Environments

In this paper, we study the optimistic online convex optimization proble...
research
09/12/2019

Nonstationary Nonparametric Online Learning: Balancing Dynamic Regret and Model Parsimony

An open challenge in supervised learning is conceptual drift: a data poi...
research
10/23/2021

Coarse-Grained Smoothness for RL in Metric Spaces

Principled decision-making in continuous state–action spaces is impossib...

Please sign up or login with your details

Forgot password? Click here to reset