The CCP Selector: Scalable Algorithms for Sparse Ridge Regression from Chance-Constrained Programming

06/11/2018
by   Weijun Xie, et al.
0

Sparse regression and variable selection for large-scale data have been rapidly developed in the past decades. This work focuses on sparse ridge regression, which considers the exact L_0 norm to pursue the sparsity. We pave out a theoretical foundation to understand why many existing approaches may not work well for this problem, in particular on large-scale datasets. Inspired by reformulating the problem as a chance-constrained program, we derive a novel mixed integer second order conic (MISOC) reformulation and prove that its continuous relaxation is equivalent to that of the convex integer formulation proposed in a recent work. Based upon these two formulations, we develop two new scalable algorithms, the greedy and randomized algorithms, for sparse ridge regression with desirable theoretical properties. The proposed algorithms are proved to yield near-optimal solutions under mild conditions. In the case of much larger dimensions, we propose to integrate the greedy algorithm with the randomized algorithm, which can greedily search the features from the nonzero subset identified by the continuous relaxation of the MISOC formulation. The merits of the proposed methods are elaborated through a set of numerical examples in comparison with several existing ones.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/29/2018

Convex Hull Formulations for Mixed-Integer Multilinear Functions

In this paper, we present convex hull formulations for a mixed-integer, ...
research
11/30/2020

A Kaczmarz Method with Simple Random Sampling for Solving Large Linear Systems

The Kaczmarz method is a popular iterative scheme for solving large-scal...
research
07/03/2019

A unified approach to mixed-integer optimization: Nonlinear formulations and scalable algorithms

We propose a unified framework to address a family of classical mixed-in...
research
07/26/2016

Uniform Approximation by Neural Networks Activated by First and Second Order Ridge Splines

We establish sup-norm error bounds for functions that are approximated b...
research
04/13/2023

OKRidge: Scalable Optimal k-Sparse Ridge Regression for Learning Dynamical Systems

We consider an important problem in scientific discovery, identifying sp...
research
06/05/2023

Compressed Sensing: A Discrete Optimization Approach

We study the Compressed Sensing (CS) problem, which is the problem of fi...
research
08/28/2020

Exact and Approximation Algorithms for Sparse PCA

Sparse PCA (SPCA) is a fundamental model in machine learning and data an...

Please sign up or login with your details

Forgot password? Click here to reset