Accelerated Message Passing for Entropy-Regularized MAP Inference

by   Jonathan N. Lee, et al.

Maximum a posteriori (MAP) inference in discrete-valued Markov random fields is a fundamental problem in machine learning that involves identifying the most likely configuration of random variables given a distribution. Due to the difficulty of this combinatorial problem, linear programming (LP) relaxations are commonly used to derive specialized message passing algorithms that are often interpreted as coordinate descent on the dual LP. To achieve more desirable computational properties, a number of methods regularize the LP with an entropy term, leading to a class of smooth message passing algorithms with convergence guarantees. In this paper, we present randomized methods for accelerating these algorithms by leveraging techniques that underlie classical accelerated gradient methods. The proposed algorithms incorporate the familiar steps of standard smooth message passing algorithms, which can be viewed as coordinate minimization steps. We show that these accelerated variants achieve faster rates for finding ϵ-optimal points of the unregularized problem, and, when the LP is tight, we prove that the proposed algorithms recover the true MAP solution in fewer iterations than standard message passing algorithms.



There are no comments yet.


page 1

page 2

page 3

page 4


An Efficient Message-Passing Algorithm for the M-Best MAP Problem

Much effort has been directed at algorithms for obtaining the highest pr...

Accelerating Message Passing for MAP with Benders Decomposition

We introduce a novel mechanism to tighten the local polytope relaxation ...

Tightening LP Relaxations for MAP using Message Passing

Linear Programming (LP) relaxations have become powerful tools for findi...

Efficient MRF Energy Minimization via Adaptive Diminishing Smoothing

We consider the linear programming relaxation of an energy minimization ...

MAP Estimation of Semi-Metric MRFs via Hierarchical Graph Cuts

We consider the task of obtaining the maximum a posteriori estimate of d...

A residual-based message passing algorithm for constraint satisfaction problems

Message passing algorithms, whose iterative nature captures well complic...

Learning Data Dependency with Communication Cost

In this paper, we consider the problem of recovering a graph that repres...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.