Stability and Generalization for Randomized Coordinate Descent

08/17/2021
by   Puyu Wang, et al.
0

Randomized coordinate descent (RCD) is a popular optimization algorithm with wide applications in solving various machine learning problems, which motivates a lot of theoretical analysis on its convergence behavior. As a comparison, there is no work studying how the models trained by RCD would generalize to test examples. In this paper, we initialize the generalization analysis of RCD by leveraging the powerful tool of algorithmic stability. We establish argument stability bounds of RCD for both convex and strongly convex objectives, from which we develop optimal generalization bounds by showing how to early-stop the algorithm to tradeoff the estimation and optimization. Our analysis shows that RCD enjoys better stability as compared to stochastic gradient descent.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2021

Stability and Generalization of Stochastic Gradient Methods for Minimax Problems

Many machine learning problems can be formulated as minimax problems suc...
research
07/07/2023

Stability and Generalization of Stochastic Compositional Gradient Descent Algorithms

Many machine learning tasks can be formulated as a stochastic compositio...
research
10/23/2017

Stability and Generalization of Learning Algorithms that Converge to Global Optima

We establish novel generalization bounds for learning algorithms that co...
research
06/14/2022

Stability and Generalization of Stochastic Optimization with Nonconvex and Nonsmooth Problems

Stochastic optimization has found wide applications in minimizing object...
research
03/30/2020

Explicit Regularization of Stochastic Gradient Methods through Duality

We consider stochastic gradient methods under the interpolation regime w...
research
10/03/2022

On Stability and Generalization of Bilevel Optimization Problem

(Stochastic) bilevel optimization is a frequently encountered problem in...
research
11/23/2021

Simple Stochastic and Online Gradient Descent Algorithms for Pairwise Learning

Pairwise learning refers to learning tasks where the loss function depen...

Please sign up or login with your details

Forgot password? Click here to reset