Explicit Regularization of Stochastic Gradient Methods through Duality

03/30/2020
by   Anant Raj, et al.
0

We consider stochastic gradient methods under the interpolation regime where a perfect fit can be obtained (minimum loss at each observation). While previous work highlighted the implicit regularization of such algorithms, we consider an explicit regularization framework as a minimum Bregman divergence convex feasibility problem. Using convex duality, we propose randomized Dykstra-style algorithms based on randomized dual coordinate ascent. For non-accelerated coordinate descent, we obtain an algorithm which bears strong similarities with (non-averaged) stochastic mirror descent on specific functions, as it is is equivalent for quadratic objectives, and equivalent in the early iterations for more general objectives. It comes with the benefit of an explicit convergence theorem to a minimum norm solution. For accelerated coordinate descent, we obtain a new algorithm that has better convergence properties than existing stochastic gradient methods in the interpolating regime. This leads to accelerated versions of the perceptron for generic ℓ_p-norm regularizers, which we illustrate in experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2018

Accelerated Randomized Coordinate Descent Methods for Stochastic Optimization and Online Learning

We propose accelerated randomized coordinate descent algorithms for stoc...
research
06/08/2015

Linear Convergence of the Randomized Feasible Descent Method Under the Weak Strong Convexity Assumption

In this paper we generalize the framework of the feasible descent method...
research
05/23/2018

Adaptive Stochastic Gradient Langevin Dynamics: Taming Convergence and Saddle Point Escape Time

In this paper, we propose a new adaptive stochastic gradient Langevin dy...
research
08/17/2021

Stability and Generalization for Randomized Coordinate Descent

Randomized coordinate descent (RCD) is a popular optimization algorithm ...
research
07/04/2023

Accelerated stochastic approximation with state-dependent noise

We consider a class of stochastic smooth convex optimization problems un...
research
08/11/2022

An Accelerated Doubly Stochastic Gradient Method with Faster Explicit Model Identification

Sparsity regularized loss minimization problems play an important role i...
research
03/26/2018

Revisiting First-Order Convex Optimization Over Linear Spaces

Two popular examples of first-order optimization methods over linear spa...

Please sign up or login with your details

Forgot password? Click here to reset