Classical algorithms and quantum limitations for maximum cut on high-girth graphs

06/10/2021 ∙ by Boaz Barak, et al. ∙ 0

We study the performance of local quantum algorithms such as the Quantum Approximate Optimization Algorithm (QAOA) for the maximum cut problem, and their relationship to that of classical algorithms. (1) We prove that every (quantum or classical) one-local algorithm achieves on D-regular graphs of girth > 5 a maximum cut of at most 1/2 + C/√(D) for C=1/√(2)≈ 0.7071. This is the first such result showing that one-local algorithms achieve a value bounded away from the true optimum for random graphs, which is 1/2 + P_*/√(D) + o(1/√(D)) for P_* ≈ 0.7632. (2) We show that there is a classical k-local algorithm that achieves a value of 1/2 + C/√(D) - O(1/√(k)) for D-regular graphs of girth > 2k+1, where C = 2/π≈ 0.6366. This is an algorithmic version of the existential bound of Lyons and is related to the algorithm of Aizenman, Lebowitz, and Ruelle (ALR) for the Sherrington-Kirkpatrick model. This bound is better than that achieved by the one-local and two-local versions of QAOA on high-girth graphs. (3) Through computational experiments, we give evidence that the ALR algorithm achieves better performance than constant-locality QAOA for random D-regular graphs, as well as other natural instances, including graphs that do have short cycles. Our experimental work suggests that it could be possible to extend beyond our theoretical constraints. This points at the tantalizing possibility that O(1)-local quantum maximum-cut algorithms might be *pointwise dominated* by polynomial-time classical algorithms, in the sense that there is a classical algorithm outputting cuts of equal or better quality *on every possible instance*. This is in contrast to the evidence that polynomial-time algorithms cannot simulate the probability distributions induced by local quantum algorithms.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 14

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.