Joint and conditional estimation of tagging and parsing models

05/07/2001
by   Mark Johnson, et al.
0

This paper compares two different ways of estimating statistical language models. Many statistical NLP tagging and parsing models are estimated by maximizing the (joint) likelihood of the fully-observed training data. However, since these applications only require the conditional probability distributions, these distributions can in principle be learnt by maximizing the conditional likelihood of the training data. Perhaps somewhat surprisingly, models estimated by maximizing the joint were superior to models estimated by maximizing the conditional, even though some of the latter models intuitively had access to "more information".

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset