Smooth minimization of nonsmooth functions with parallel coordinate descent methods

09/23/2013
by   Olivier Fercoq, et al.
0

We study the performance of a family of randomized parallel coordinate descent methods for minimizing the sum of a nonsmooth and separable convex functions. The problem class includes as a special case L1-regularized L1 regression and the minimization of the exponential loss ("AdaBoost problem"). We assume the input data defining the loss function is contained in a sparse m× n matrix A with at most ω nonzeros in each row. Our methods need O(n β/τ) iterations to find an approximate solution with high probability, where τ is the number of processors and β = 1 + (ω-1)(τ-1)/(n-1) for the fastest variant. The notation hides dependence on quantities such as the required accuracy and confidence levels and the distance of the starting iterate from an optimal point. Since β/τ is a decreasing function of τ, the method needs fewer iterations when more processors are used. Certain variants of our algorithms perform on average only O((A)/n) arithmetic operations during a single iteration per processor and, because β decreases when ω does, fewer iterations are needed for sparser problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/04/2012

Parallel Coordinate Descent Methods for Big Data Optimization

In this work we show that randomized (block) coordinate descent methods ...
research
12/20/2013

Accelerated, Parallel and Proximal Coordinate Descent

We propose a new stochastic coordinate descent method for minimizing the...
research
07/14/2011

Iteration Complexity of Randomized Block-Coordinate Descent Methods for Minimizing a Composite Function

In this paper we develop a randomized block-coordinate descent method fo...
research
02/09/2015

Random Coordinate Descent Methods for Minimizing Decomposable Submodular Functions

Submodular function minimization is a fundamental optimization problem t...
research
10/08/2013

Distributed Coordinate Descent Method for Learning with Big Data

In this paper we develop and analyze Hydra: HYbriD cooRdinAte descent me...
research
08/06/2021

Adjusting PageRank parameters and Comparing results

The effect of adjusting damping factor α and tolerance τ on iterations n...
research
08/30/2010

Fixed-point and coordinate descent algorithms for regularized kernel methods

In this paper, we study two general classes of optimization algorithms f...

Please sign up or login with your details

Forgot password? Click here to reset