Convergence of Online Mirror Descent Algorithms

02/18/2018
by   Yunwen Lei, et al.
0

In this paper we consider online mirror descent (OMD) algorithms, a class of scalable online learning algorithms exploiting data geometric structures through mirror maps. Necessary and sufficient conditions are presented in terms of the step size sequence {η_t}_t for the convergence of an OMD algorithm with respect to the expected Bregman distance induced by the mirror map. The condition is _t→∞η_t=0, ∑_t=1^∞η_t=∞ in the case of positive variances. It is reduced to ∑_t=1^∞η_t=∞ in the case of zero variances for which the linear convergence may be achieved by taking a constant step size sequence. A sufficient condition on the almost sure convergence is also given. We establish tight error bounds under mild conditions on the mirror map, the loss function, and the regularizer. Our results are achieved by some novel analysis on the one-step progress of the OMD algorithm using smoothness and strong convexity of the mirror map and the loss function.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/02/2015

Unregularized Online Learning Algorithms with General Loss Functions

In this paper, we consider unregularized online learning algorithms in a...
research
05/28/2019

Concavifiability and convergence: necessary and sufficient conditions for gradient descent analysis

Convergence of the gradient descent algorithm has been attracting renewe...
research
11/24/2022

Online Regularized Learning Algorithm for Functional Data

In recent years, functional linear models have attracted growing attenti...
research
11/21/2021

A Data-Driven Line Search Rule for Support Recovery in High-dimensional Data Analysis

In this work, we consider the algorithm to the (nonlinear) regression pr...
research
09/15/2023

Convergence of ADAM with Constant Step Size in Non-Convex Settings: A Simple Proof

In neural network training, RMSProp and ADAM remain widely favoured opti...
research
12/13/2022

Self-adaptive algorithms for quasiconvex programming and applications to machine learning

For solving a broad class of nonconvex programming problems on an unboun...
research
10/22/2020

Tight two-level convergence of Linear Parareal and MGRIT: Extensions and implications in practice

Two of the most popular parallel-in-time methods are Parareal and multig...

Please sign up or login with your details

Forgot password? Click here to reset