DeepAI AI Chat
Log In Sign Up

Inference in Graphical Models via Semidefinite Programming Hierarchies

by   Murat A. Erdogdu, et al.
Stanford University

Maximum A posteriori Probability (MAP) inference in graphical models amounts to solving a graph-structured combinatorial optimization problem. Popular inference algorithms such as belief propagation (BP) and generalized belief propagation (GBP) are intimately related to linear programming (LP) relaxation within the Sherali-Adams hierarchy. Despite the popularity of these algorithms, it is well understood that the Sum-of-Squares (SOS) hierarchy based on semidefinite programming (SDP) can provide superior guarantees. Unfortunately, SOS relaxations for a graph with n vertices require solving an SDP with n^Θ(d) variables where d is the degree in the hierarchy. In practice, for d> 4, this approach does not scale beyond a few tens of variables. In this paper, we propose binary SDP relaxations for MAP inference using the SOS hierarchy with two innovations focused on computational efficiency. Firstly, in analogy to BP and its variants, we only introduce decision variables corresponding to contiguous regions in the graphical model. Secondly, we solve the resulting SDP using a non-convex Burer-Monteiro style method, and develop a sequential rounding procedure. We demonstrate that the resulting algorithm can solve problems with tens of thousands of variables within minutes, and outperforms BP and GBP on practical problems such as image denoising and Ising spin glasses. Finally, for specific graph types, we establish a sufficient condition for the tightness of the proposed partial SOS relaxation.


page 7

page 12


Belief Propagation for Linear Programming

Belief Propagation (BP) is a popular, distributed heuristic for performi...

Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization

Maximum a posteriori (MAP) inference is a fundamental computational para...

Efficient semidefinite bounds for multi-label discrete graphical models

By concisely representing a joint function of many variables as the comb...

Minimum Weight Perfect Matching via Blossom Belief Propagation

Max-product Belief Propagation (BP) is a popular message-passing algorit...

A Generalized Loop Correction Method for Approximate Inference in Graphical Models

Belief Propagation (BP) is one of the most popular methods for inference...

Scalable Semidefinite Relaxation for Maximum A Posterior Estimation

Maximum a posteriori (MAP) inference over discrete Markov random fields ...

The p-convolution forest: a method for solving graphical models with additive probabilistic equations

Convolution trees, loopy belief propagation, and fast numerical p-convol...