Feature Clustering for Accelerating Parallel Coordinate Descent

12/17/2012
by   Chad Scherrer, et al.
0

Large-scale L1-regularized loss minimization problems arise in high-dimensional applications such as compressed sensing and high-dimensional supervised learning, including classification and regression problems. High-performance algorithms and implementations are critical to efficiently solving these problems. Building upon previous work on coordinate descent algorithms for L1-regularized problems, we introduce a novel family of algorithms called block-greedy coordinate descent that includes, as special cases, several existing algorithms such as SCD, Greedy CD, Shotgun, and Thread-Greedy. We give a unified convergence analysis for the family of block-greedy algorithms. The analysis suggests that block-greedy coordinate descent can better exploit parallelism if features are clustered so that the maximum inner product between features in different blocks is small. Our theoretical convergence analysis is supported with experimental re- sults using data from diverse real-world applications. We hope that algorithmic approaches and convergence analysis we provide will not only advance the field, but will also encourage researchers to systematically explore the design space of algorithms for solving large-scale L1-regularization problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/27/2012

Scaling Up Coordinate Descent Algorithms for Large ℓ_1 Regularization Problems

We present a generic framework for parallel coordinate descent (CD) algo...
research
10/16/2018

Efficient Greedy Coordinate Descent for Composite Problems

Coordinate descent with random coordinate selection is the current state...
research
03/04/2022

Greedy double subspaces coordinate descent method via orthogonalization

The coordinate descent method is an effective iterative method for solvi...
research
12/24/2020

Stochastic Steepest Descent Methods for Linear Systems: Greedy Sampling Momentum

Recently proposed adaptive Sketch Project (SP) methods connect sever...
research
11/13/2016

Accelerated Variance Reduced Block Coordinate Descent

Algorithms with fast convergence, small number of data access, and low p...
research
10/07/2013

Parallel coordinate descent for the Adaboost problem

We design a randomised parallel version of Adaboost based on previous st...
research
02/17/2016

Large Scale Kernel Learning using Block Coordinate Descent

We demonstrate that distributed block coordinate descent can quickly sol...

Please sign up or login with your details

Forgot password? Click here to reset