The power of sum-of-squares for detecting hidden structures

10/13/2017
by   Samuel B. Hopkins, et al.
0

We study planted problems---finding hidden structures in random noisy inputs---through the lens of the sum-of-squares semidefinite programming hierarchy (SoS). This family of powerful semidefinite programs has recently yielded many new algorithms for planted problems, often achieving the best known polynomial-time guarantees in terms of accuracy of recovered solutions and robustness to noise. One theme in recent work is the design of spectral algorithms which match the guarantees of SoS algorithms for planted problems. Classical spectral algorithms are often unable to accomplish this: the twist in these new spectral algorithms is the use of spectral structure of matrices whose entries are low-degree polynomials of the input variables. We prove that for a wide class of planted problems, including refuting random constraint satisfaction problems, tensor and sparse PCA, densest-k-subgraph, community detection in stochastic block models, planted clique, and others, eigenvalues of degree-d matrix polynomials are as powerful as SoS semidefinite programs of roughly degree d. For such problems it is therefore always possible to match the guarantees of SoS without solving a large semidefinite program. Using related ideas on SoS algorithms and low-degree matrix polynomials (and inspired by recent work on SoS and the planted clique problem by Barak et al.), we prove new nearly-tight SoS lower bounds for the tensor and sparse principal component analysis problems. Our lower bounds for sparse principal component analysis are the first to suggest that going beyond existing algorithms for this problem may require sub-exponential time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/09/2020

Machinery for Proving Sum-of-Squares Lower Bounds on Certification Problems

In this paper, we construct general machinery for proving Sum-of-Squares...
research
06/27/2017

Fast and robust tensor decomposition with applications to dictionary learning

We develop fast spectral algorithms for tensor decomposition that match ...
research
11/12/2020

Sparse PCA: Algorithms, Adversarial Perturbations and Certificates

We study efficient algorithms for Sparse PCA in standard statistical mod...
research
02/09/2023

Nonlinear Random Matrices and Applications to the Sum of Squares Hierarchy

We develop new tools in the theory of nonlinear random matrices and appl...
research
09/06/2022

Concentration of polynomial random matrices via Efron-Stein inequalities

Analyzing concentration of large random matrices is a common task in a w...
research
04/08/2019

The Kikuchi Hierarchy and Tensor PCA

For the tensor PCA (principal component analysis) problem, we propose a ...
research
12/03/2020

Average-Case Integrality Gap for Non-Negative Principal Component Analysis

Montanari and Richard (2015) asked whether a natural semidefinite progra...

Please sign up or login with your details

Forgot password? Click here to reset