Revisiting Frank-Wolfe for Polytopes: Strict Complementary and Sparsity

05/31/2020
by   Dan Garber, et al.
0

In recent years it was proved that simple modifications of the classical Frank-Wolfe algorithm (aka conditional gradient algorithm) for smooth convex minimization over convex and compact polytopes, converge with linear rate, assuming the objective function has the quadratic growth property. However, the rate of these methods depends explicitly on the dimension of the problem which cannot explain their empirical success for large scale problems. In this paper we first demonstrate that already for very simple problems and even when the optimal solution lies on a low-dimensional face of the polytope, such dependence on the dimension cannot be avoided in worst case. We then revisit the addition of a strict complementary assumption already considered in Wolfe's classical book <cit.>, and prove that under this condition, the Frank-Wolfe method with away-steps and line-search converges linearly with rate that depends explicitly only on the dimension of the optimal face, hence providing a significant improvement in case the optimal solution is sparse. We motivate this strict complementary condition by proving that it implies sparsity-robustness of optimal solutions to noise.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/03/2021

Frank-Wolfe with a Nearest Extreme Point Oracle

We consider variants of the classical Frank-Wolfe algorithm for constrai...
research
08/03/2023

Efficiency of First-Order Methods for Low-Rank Tensor Recovery with the Tensor Nuclear Norm Under Strict Complementarity

We consider convex relaxations for recovering low-rank tensors based on ...
research
02/08/2022

Efficient Algorithms for High-Dimensional Convex Subspace Optimization via Strict Complementarity

We consider optimization problems in which the goal is find a k-dimensio...
research
11/24/2015

Generalized Conjugate Gradient Methods for ℓ_1 Regularized Convex Quadratic Programming with Finite Convergence

The conjugate gradient (CG) method is an efficient iterative method for ...
research
03/02/2021

Convergence Rate of the (1+1)-Evolution Strategy with Success-Based Step-Size Adaptation on Convex Quadratic Functions

The (1+1)-evolution strategy (ES) with success-based step-size adaptatio...
research
05/26/2019

Partial minimization of strict convex functions and tensor scaling

Assume that f is a strict convex function with a unique minimum in R^n. ...
research
02/12/2016

Scale-free network optimization: foundations and algorithms

We investigate the fundamental principles that drive the development of ...

Please sign up or login with your details

Forgot password? Click here to reset