Optimal Average-Case Reductions to Sparse PCA: From Weak Assumptions to Strong Hardness
In the past decade, sparse principal component analysis has emerged as an archetypal problem for illustrating statistical-computational tradeoffs. This trend has largely been driven by a line of research aiming to characterize the average-case complexity of sparse PCA through reductions from the planted clique (PC) conjecture - which conjectures that there is no polynomial-time algorithm to detect a planted clique of size K = o(N^1/2) in G(N, 1/2). All previous reductions to sparse PCA either fail to show tight computational lower bounds matching existing algorithms or show lower bounds for formulations of sparse PCA other than its canonical generative model, the spiked covariance model. Also, these lower bounds all quickly degrade with the exponent in the PC conjecture. Specifically, when only given the PC conjecture up to K = o(N^α) where α < 1/2, there is no sparsity level k at which these lower bounds remain tight. If α< 1/3 these reductions fail to even show the existence of a statistical-computational tradeoff at any sparsity k. We give a reduction from PC that yields the first full characterization of the computational barrier in the spiked covariance model, providing tight lower bounds at all sparsities k. We also show the surprising result that weaker forms of the PC conjecture up to clique size K = o(N^α) for any given α∈ (0, 1/2] imply tight computational lower bounds for sparse PCA at sparsities k = o(n^α/3). This shows that even a mild improvement in the signal strength needed by the best known polynomial-time sparse PCA algorithms would imply that the hardness threshold for PC is subpolynomial. This is the first instance of a suboptimal hardness assumption implying optimal lower bounds for another problem in unsupervised learning.
READ FULL TEXT