Maximum Likelihood Learning With Arbitrary Treewidth via Fast-Mixing Parameter Sets

09/30/2015
by   Justin Domke, et al.
0

Inference is typically intractable in high-treewidth undirected graphical models, making maximum likelihood learning a challenge. One way to overcome this is to restrict parameters to a tractable set, most typically the set of tree-structured parameters. This paper explores an alternative notion of a tractable set, namely a set of "fast-mixing parameters" where Markov chain Monte Carlo (MCMC) inference can be guaranteed to quickly converge to the stationary distribution. While it is common in practice to approximate the likelihood gradient using samples obtained from MCMC, such procedures lack theoretical guarantees. This paper proves that for any exponential family with bounded sufficient statistics, (not just graphical models) when parameters are constrained to a fast-mixing set, gradient descent with gradients approximated by sampling will approximate the maximum likelihood solution inside the set with high-probability. When unregularized, to find a solution epsilon-accurate in log-likelihood requires a total amount of effort cubic in 1/epsilon, disregarding logarithmic factors. When ridge-regularized, strong convexity allows a solution epsilon-accurate in parameter distance with effort quadratic in 1/epsilon. Both of these provide of a fully-polynomial time randomized approximation scheme.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/03/2014

Projecting Ising Model Parameters for Fast Mixing

Inference in general Ising models is difficult, due to high treewidth ma...
research
07/11/2012

Bayesian Learning in Undirected Graphical Models: Approximate MCMC algorithms

Bayesian learning in undirected graphical models|computing posterior dis...
research
11/05/2014

Projecting Markov Random Field Parameters for Fast Mixing

Markov chain Monte Carlo (MCMC) algorithms are simple and extremely powe...
research
07/18/2019

A Polynomial Time Algorithm for Log-Concave Maximum Likelihood via Locally Exponential Families

We consider the problem of computing the maximum likelihood multivariate...
research
12/27/2019

Worst-Case Polynomial-Time Exact MAP Inference on Discrete Models with Global Dependencies

Considering the worst-case scenario, junction tree algorithm remains the...
research
02/14/2012

What Cannot be Learned with Bethe Approximations

We address the problem of learning the parameters in graphical models wh...
research
02/21/2019

Approximate selective inference via maximum likelihood

We consider an approximate version of the conditional approach to select...

Please sign up or login with your details

Forgot password? Click here to reset