Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization

02/27/2022
by   Idan Amir, et al.
0

We consider linear prediction with a convex Lipschitz loss, or more generally, stochastic convex optimization problems of generalized linear form, i.e. where each instantaneous loss is a scalar convex function of a linear function. We show that in this setting, early stopped Gradient Descent (GD), without any explicit regularization or projection, ensures excess error at most ϵ (compared to the best possible with unit Euclidean norm) with an optimal, up to logarithmic factors, sample complexity of Õ(1/ϵ^2) and only Õ(1/ϵ^2) iterations. This contrasts with general stochastic convex optimization, where Ω(1/ϵ^4) iterations are needed Amir et al. [2021b]. The lower iteration complexity is ensured by leveraging uniform convergence rather than stability. But instead of uniform convergence in a norm ball, which we show can guarantee suboptimal learning using Θ(1/ϵ^4) samples, we rely on uniform convergence in a distribution-dependent ball.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/29/2021

Never Go Full Batch (in Stochastic Convex Optimization)

We study the generalization performance of full-batch optimization algor...
research
06/02/2023

Convex and Non-Convex Optimization under Generalized Smoothness

Classical analysis of convex and non-convex optimization methods often r...
research
11/10/2021

Efficient Projection-Free Online Convex Optimization with Membership Oracle

In constrained convex optimization, existing methods based on the ellips...
research
05/31/2023

Parameter-free projected gradient descent

We consider the problem of minimizing a convex function over a closed co...
research
11/02/2022

Quasi-Newton Steps for Efficient Online Exp-Concave Optimization

The aim of this paper is to design computationally-efficient and optimal...
research
06/10/2020

On Uniform Convergence and Low-Norm Interpolation Learning

We consider an underdetermined noisy linear regression model where the m...
research
05/15/2023

Convex optimization over a probability simplex

We propose a new iteration scheme, the Cauchy-Simplex, to optimize conve...

Please sign up or login with your details

Forgot password? Click here to reset