Viterbi training in PRISM

03/22/2013
by   Taisuke Sato, et al.
0

VT (Viterbi training), or hard EM, is an efficient way of parameter learning for probabilistic models with hidden variables. Given an observation y, it searches for a state of hidden variables x that maximizes p(x,y |θ) by coordinate ascent on parameters θ and x. In this paper we introduce VT to PRISM, a logic-based probabilistic modeling system for generative models. VT improves PRISM in three ways. First VT in PRISM converges faster than EM in PRISM due to the VT's termination condition. Second, parameters learned by VT often show good prediction performance compared to those learned by EM. We conducted two parsing experiments with probabilistic grammars while learning parameters by a variety of inference methods, i.e. VT, EM, MAP and VB. The result is that VT achieved the best parsing accuracy among them in both experiments. Also we conducted a similar experiment for classification tasks where a hidden variable is not a prediction target unlike probabilistic grammars. We found that in such a case VT does not necessarily yield superior performance. Third since VT always deals with a single probability of a single explanation, Viterbi explanation, the exclusiveness condition that is imposed on PRISM programs is no more required if we learn parameters by VT. Last but not least we can say that as VT in PRISM is general and applicable to any PRISM program, it largely reduces the need for the user to develop a specific VT algorithm for a specific model. Furthermore since VT in PRISM can be used just by setting a PRISM flag appropriately, it makes VT easily accessible to (probabilistic) logic programmers. To appear in Theory and Practice of Logic Programming (TPLP).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2022

Probabilistic Systems with Hidden State and Unobservable Transitions

We consider probabilistic systems with hidden state and unobservable tra...
research
01/23/2013

Accelerating EM: An Empirical Study

Many applications require that we learn the parameters of a model from d...
research
03/20/2013

Integrating Probabilistic Rules into Neural Networks: A Stochastic EM Learning Algorithm

The EM-algorithm is a general procedure to get maximum likelihood estima...
research
08/04/2020

MAP Inference for Probabilistic Logic Programming

In Probabilistic Logic Programming (PLP) the most commonly studied infer...
research
04/25/2013

Inference and learning in probabilistic logic programs using weighted Boolean formulas

Probabilistic logic programs are logic programs in which some of the fac...
research
06/09/2011

Parameter Learning of Logic Programs for Symbolic-Statistical Modeling

We propose a logical/mathematical framework for statistical parameter le...

Please sign up or login with your details

Forgot password? Click here to reset