DeepAI AI Chat
Log In Sign Up

Message-Passing Algorithms for Quadratic Programming Formulations of MAP Estimation

by   Akshat Kumar, et al.

Computing maximum a posteriori (MAP) estimation in graphical models is an important inference problem with many applications. We present message-passing algorithms for quadratic programming (QP) formulations of MAP estimation for pairwise Markov random fields. In particular, we use the concave-convex procedure (CCCP) to obtain a locally optimal algorithm for the non-convex QP formulation. A similar technique is used to derive a globally convergent algorithm for the convex QP relaxation of MAP. We also show that a recently developed expectation-maximization (EM) algorithm for the QP formulation of MAP can be derived from the CCCP perspective. Experiments on synthetic and real-world problems confirm that our new approach is competitive with max-product and its variations. Compared with CPLEX, we achieve more than an order-of-magnitude speedup in solving optimally the convex QP relaxation.


MAP Estimation, Message Passing, and Perfect Graphs

Efficiently finding the maximum a posteriori (MAP) configuration of a gr...

An Efficient Message-Passing Algorithm for the M-Best MAP Problem

Much effort has been directed at algorithms for obtaining the highest pr...

MAP Estimation of Semi-Metric MRFs via Hierarchical Graph Cuts

We consider the task of obtaining the maximum a posteriori estimate of d...

Accelerating Message Passing for MAP with Benders Decomposition

We introduce a novel mechanism to tighten the local polytope relaxation ...

Tightening LP Relaxations for MAP using Message Passing

Linear Programming (LP) relaxations have become powerful tools for findi...

Sparse Multinomial Logistic Regression via Approximate Message Passing

For the problem of multi-class linear classification and feature selecti...