Efficient Algorithms for High-Dimensional Convex Subspace Optimization via Strict Complementarity

02/08/2022
by   Dan Garber, et al.
0

We consider optimization problems in which the goal is find a k-dimensional subspace of ^n, k<<n, which minimizes a convex and smooth loss. Such problemsgeneralize the fundamental task of principal component analysis (PCA) to include robust and sparse counterparts, and logistic PCA for binary data, among others. While this problem is not convex it admits natural algorithms with very efficient iterations and memory requirements, which is highly desired in high-dimensional regimes however, arguing about their fast convergence to a global optimal solution is difficult. On the other hand, there exists a simple convex relaxation for which convergence to the global optimum is straightforward, however corresponding algorithms are not efficient when the dimension is very large. In this work we present a natural deterministic sufficient condition so that the optimal solution to the convex relaxation is unique and is also the optimal solution to the original nonconvex problem. Mainly, we prove that under this condition, a natural highly-efficient nonconvex gradient method, which we refer to as gradient orthogonal iteration, when initialized with a "warm-start", converges linearly for the nonconvex problem. We also establish similar results for the nonconvex projected gradient method, and the Frank-Wolfe method when applied to the convex relaxation. We conclude with empirical evidence on synthetic data which demonstrate the appeal of our approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/07/2019

Stochastic Approximation Algorithms for Principal Component Analysis

Principal Component Analysis is a novel way of of dimensionality reducti...
research
12/03/2019

Linear Convergence of Frank-Wolfe for Rank-One Matrix Recovery Without Strong Convexity

We consider convex optimization problems which are widely used as convex...
research
12/30/2018

Exact Guarantees on the Absence of Spurious Local Minima for Non-negative Robust Principal Component Analysis

This work is concerned with the non-negative robust principal component ...
research
05/31/2020

Revisiting Frank-Wolfe for Polytopes: Strict Complementary and Sparsity

In recent years it was proved that simple modifications of the classical...
research
12/24/2018

Dual Principal Component Pursuit: Probability Analysis and Efficient Algorithms

Recent methods for learning a linear subspace from data corrupted by out...
research
05/21/2018

A Nonconvex Projection Method for Robust PCA

Robust principal component analysis (RPCA) is a well-studied problem wit...
research
08/04/2020

Convex and Nonconvex Optimization Are Both Minimax-Optimal for Noisy Blind Deconvolution

We investigate the effectiveness of convex relaxation and nonconvex opti...

Please sign up or login with your details

Forgot password? Click here to reset