Stochastic optimization and sparse statistical recovery: An optimal algorithm for high dimensions

07/18/2012
by   Alekh Agarwal, et al.
0

We develop and analyze stochastic optimization algorithms for problems in which the expected loss is strongly convex, and the optimum is (approximately) sparse. Previous approaches are able to exploit only one of these two structures, yielding an (/T) convergence rate for strongly convex objectives in dimensions, and an (√(()/T)) convergence rate when the optimum is -sparse. Our algorithm is based on successively solving a series of ℓ_1-regularized optimization problems using Nesterov's dual averaging algorithm. We establish that the error of our solution after T iterations is at most (()/T), with natural extensions to approximate sparsity. Our results apply to locally Lipschitz losses including the logistic, exponential, hinge and least-squares losses. By recourse to statistical minimax results, we show that our convergence rates are optimal up to multiplicative constant factors. The effectiveness of our approach is also confirmed in numerical simulations, in which we compare to several baselines on a least-squares regression problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/11/2017

Dynamic Stochastic Approximation for Multi-stage Stochastic Optimization

In this paper, we consider multi-stage stochastic optimization problems ...
research
03/22/2011

Randomized Smoothing for Stochastic Optimization

We analyze convergence rates of stochastic optimization procedures for n...
research
01/08/2018

How To Make the Gradients Small Stochastically

In convex stochastic optimization, convergence rates in terms of minimiz...
research
10/11/2011

The Generalization Ability of Online Algorithms for Dependent Data

We study the generalization performance of online learning algorithms tr...
research
02/25/2020

Can speed up the convergence rate of stochastic gradient methods to O(1/k^2) by a gradient averaging strategy?

In this paper we consider the question of whether it is possible to appl...
research
02/03/2021

Query Complexity of Least Absolute Deviation Regression via Robust Uniform Convergence

Consider a regression problem where the learner is given a large collect...
research
10/18/2016

Analysis and Implementation of an Asynchronous Optimization Algorithm for the Parameter Server

This paper presents an asynchronous incremental aggregated gradient algo...

Please sign up or login with your details

Forgot password? Click here to reset