Dropping Convexity for Faster Semi-definite Optimization

09/14/2015
by   Srinadh Bhojanapalli, et al.
0

We study the minimization of a convex function f(X) over the set of n× n positive semi-definite matrices, but when the problem is recast as _U g(U) := f(UU^), with U ∈R^n × r and r ≤ n. We study the performance of gradient descent on g---which we refer to as Factored Gradient Descent (FGD)---under standard assumptions on the original function f. We provide a rule for selecting the step size and, with this choice, show that the local convergence rate of FGD mirrors that of standard gradient descent on the original f: i.e., after k steps, the error is O(1/k) for smooth f, and exponentially small in k when f is (restricted) strongly convex. In addition, we provide a procedure to initialize FGD for (restricted) strongly convex objectives and when one only has access to f via a first-order oracle; for several problem instances, such proper initialization leads to global convergence guarantees. FGD and similar procedures are widely used in practice for problems that can be posed as matrix factorization. To the best of our knowledge, this is the first paper to provide precise convergence rate guarantees for general convex functions under standard convex assumptions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

Fast Rates for Online Gradient Descent Without Strong Convexity via Hoffman's Bound

Hoffman's classical result gives a bound on the distance of a point from...
research
05/29/2021

On Centralized and Distributed Mirror Descent: Exponential Convergence Analysis Using Quadratic Constraints

Mirror descent (MD) is a powerful first-order optimization technique tha...
research
06/10/2020

Random Reshuffling: Simple Analysis with Vast Improvements

Random Reshuffling (RR) is an algorithm for minimizing finite-sum functi...
research
03/22/2022

Local Stochastic Factored Gradient Descent for Distributed Quantum State Tomography

We propose a distributed Quantum State Tomography (QST) protocol, named ...
research
07/31/2023

Line Search for Convex Minimization

Golden-section search and bisection search are the two main principled a...
research
02/11/2020

Self-concordant analysis of Frank-Wolfe algorithms

Projection-free optimization via different variants of the Frank-Wolfe (...
research
09/29/2022

Restricted Strong Convexity of Deep Learning Models with Smooth Activations

We consider the problem of optimization of deep learning models with smo...

Please sign up or login with your details

Forgot password? Click here to reset