Linear Convergence of a Frank-Wolfe Type Algorithm over Trace-Norm Balls

08/07/2017
by   Zeyuan Allen-Zhu, et al.
0

We propose a rank-k variant of the classical Frank-Wolfe algorithm to solve convex optimization over a trace-norm ball. Our algorithm replaces the top singular-vector computation (1-SVD) in Frank-Wolfe with a top-k singular-vector computation (k-SVD), which can be done by repeatedly applying 1-SVD k times. Alternatively, our algorithm can be viewed as a rank-k restricted version of projected gradient descent. We show that our algorithm has a linear convergence rate when the objective function is smooth and strongly convex, and the optimal solution has rank at most k. This improves the convergence rate and the total time complexity of the Frank-Wolfe method and its variants.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2023

Efficiency of First-Order Methods for Low-Rank Tensor Recovery with the Tensor Nuclear Norm Under Strict Complementarity

We consider convex relaxations for recovering low-rank tensors based on ...
research
05/11/2020

On Radial Isotropic Position: Theory and Algorithms

We review the theory of, and develop algorithms for transforming a finit...
research
04/03/2021

Convergence Analysis of the Rank-Restricted Soft SVD Algorithm

The soft SVD is a robust matrix decomposition algorithm and a key compon...
research
12/03/2019

Linear Convergence of Frank-Wolfe for Rank-One Matrix Recovery Without Strong Convexity

We consider convex optimization problems which are widely used as convex...
research
12/18/2020

On the Efficient Implementation of the Matrix Exponentiated Gradient Algorithm for Low-Rank Matrix Optimization

Convex optimization over the spectrahedron, i.e., the set of all real n×...
research
08/31/2020

Super-linear convergence in the p-adic QR-algorithm

The QR-algorithm is one of the most important algorithms in linear algeb...

Please sign up or login with your details

Forgot password? Click here to reset