A Random Block-Coordinate Douglas-Rachford Splitting Method with Low Computational Complexity for Binary Logistic Regression

by   Luis M. Briceno-Arias, et al.

In this paper, we propose a new optimization algorithm for sparse logistic regression based on a stochastic version of the Douglas-Rachford splitting method. Our algorithm sweeps the training set by randomly selecting a mini-batch of data at each iteration, and it allows us to update the variables in a block coordinate manner. Our approach leverages the proximity operator of the logistic loss, which is expressed with the generalized Lambert W function. Experiments carried out on standard datasets demonstrate the efficiency of our approach w.r.t. stochastic gradient-like methods.


page 1

page 2

page 3

page 4


Multinomial Logistic Regression Algorithms via Quadratic Gradient

Multinomial logistic regression, also known by other names such as multi...

A Hidden Variables Approach to Multilabel Logistic Regression

Multilabel classification is an important problem in a wide range of dom...

Privacy-Preserving Logistic Regression Training with a Faster Gradient Variant

Logistic regression training on an encrypted dataset has been an attract...

A New Smooth Approximation to the Zero One Loss with a Probabilistic Interpretation

We examine a new form of smooth approximation to the zero one loss in wh...

Low Complexity Approximate Bayesian Logistic Regression for Sparse Online Learning

Theoretical results show that Bayesian methods can achieve lower bounds ...

Accelerated Proximal Stochastic Dual Coordinate Ascent for Regularized Loss Minimization

We introduce a proximal version of the stochastic dual coordinate ascent...

Block-Cyclic Stochastic Coordinate Descent for Deep Neural Networks

We present a stochastic first-order optimization algorithm, named BCSC, ...

Please sign up or login with your details

Forgot password? Click here to reset