 # Latent tree models

Latent tree models are graphical models defined on trees, in which only a subset of variables is observed. They were first discussed by Judea Pearl as tree-decomposable distributions to generalise star-decomposable distributions such as the latent class model. Latent tree models, or their submodels, are widely used in: phylogenetic analysis, network tomography, computer vision, causal modeling, and data clustering. They also contain other well-known classes of models like hidden Markov models, Brownian motion tree model, the Ising model on a tree, and many popular models used in phylogenetics. This article offers a concise introduction to the theory of latent tree models. We emphasise the role of tree metrics in the structural description of this model class, in designing learning algorithms, and in understanding fundamental limits of what and when can be learned.

## Authors

##### This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

## 1 Basics

In this section we define latent tree models and provide motivation to work with this model class. We present Gaussian and general Markov models as subclasses of latent tree models that admits tractable and rigorous analysis.

### 1.1 Definitions

A tree is an undirected graph without cycles. A leaf of is a vertex of degree one, an internal vertex is a vertex which is not a leaf, and an inner edge is an edge whose both ends are internal vertices. Given a tree define a rooted tree as a directed graph obtained from by picking one of its vertices and directing all edges away from . The vertex is called the root. Trees will be always leaf-labeled with the labelling set , where is the number of leaves. An undirected tree is trivalent if each internal vertex has degree precisely three. A rooted tree is a binary rooted tree if each internal vertex has precisely two children. In many applications rooted trees are depicted without using arrows, where direction is made implicit by drawing the root on the top and the leaves on the bottom; see Figure 1(c). Two special types of undirected trees are: a star tree with one internal vertex and a trivalent tree on four leaves called a quartet tree; see Figure 1(a) and (b). A forest is a collection of trees. Forests here are also leaf-labeled with the labelling set is , which means that each tree in this collection is leaf-labeled and the corresponding collection of labelling sets forms a set partition of . We define three graph operations on trees (forests). Removing an edge means removing that edge from the edge set. Contracting an edge means removing from the vertex set, adding a new vertex and edges such that is adjacent to all vertices which were adjacent to or . Suppressing a vertex of degree two means removing that vertex and replacing the two edges incident to that vertex by a single edge. Figure 1: (a) An undirected star tree with five leaves, (b) a quartet tree, (c) a binary rooted tree.

A latent tree model is defined as follows. Let

be a random vector with coordinates indexed by the vertices

of a tree and with values in the state space . Suppose that the density of , denoted by , lies in the graphical model over . This means that factorises acording to the tree

 p(y)=∏u−vΨuv(yu,yv),

where for each edge of the function is a nonnegative function. We call such a model a fully-observed tree model. Denote the set of leaves by and let . Write , , , and . The latent tree model is the family of marginal distributions of over the leaves :

 pW(x):=∫Hp(x,h)dh. (1)

In other words, the internal vertices represent unobserved random variables. The above definition extends to situations where some internal vertices are also observed. However, these seemingly more general situation does not lead to any new family of distributions; for Gaussian and general Markov models this is shown in Theorem

3.

Consider now a Bayesian network on a rooted tree obtained from

. This is a model of distributions that factorise as follows

 p(y)=pr(yr)∏u→vpv|u(yv|yu)for all % y∈Y. (2)

By standard results on Markov equivalence of directed acyclic graphs [40, 89], this model coincides with the latent tree model on for any choice of the root location. The fully-observed tree model is then fully characterized by the root distribution and by the conditional distributions for all edges in the rooted tree. The parameterisation of the corresponding latent tree model is induced by taking the margin over as in (1).

Although the definition of latent tree models is fairly general, here we make additional assumptions on the state space and possible distributions. If is a finite set then we call the corresponding latent tree model discrete. Gaussian latent tree models are latent tree models for which the vector is jointly Gaussian. In the Gaussian case we typically assume that the mean of is known and equal to zero. Mixed cases when is finite and is Gaussian conditionally on are also popular.

The parameters of latent tree models are of two kinds. The underlying tree is the discrete parameter. The continuous parameter is the parameter specifying the root distribution and the conditional distributions for each edge. To make the parameters explicit, we write and . Formally, the state space of the unobserved part of the vector is also a parameter of a latent tree model. In certain applications finding can be important; see [61, Section 3.2.4]. Here we always assume that is fixed. With this convention, there are three main learning tasks related to latent tree models:

1. Given a sample of size

from a latent tree model estimate the underlying tree

.

2. Given a sample of size from a latent tree model on a fixed tree estimate the continuous parameter .

3. Given a fully specified latent tree model and a single sample at the observed vertices, infer the states at the unobserved vertices.

There are several fundamental questions related to the first two learning tasks that we are going to address in this exposition. We will not discuss here the learning task (L3). If the model is completely specified then, by (2), we have access to the full distribution over . In this case the learning task (L3) reduces to the sum-product algorithm discussed, for example, in [10, Section 8.4].

### 1.2 Motivation and applications

Latent tree models form the most tractable family of Bayesian networks with unobserved variables, which can be used to model dependence structures when unobserved confounders are expected; see, for example, [65, Section 2] or . However, there are several other reasons why latent tree models become popular across sciences. We distinguish three main types of applications.

First, latent tree models represent a larger family of probability distributions than fully-observed tree models but retain some of their computational advantages, which is particularly important in high-dimensional settings. From (

2) it is clear that having an estimator of model parameters we obtain an estimator of the fully observed distribution and so we can very efficiently compute various marginal distributions in the model using the sum-product algorithm; see, for example, [10, Section 8.4]. Using the max-product algorithm it is also possible to efficiently infer the unobserved states.

Second, a rooted tree can represent evolutionary processes with the root representing the common ancestor and the leaves representing extant species. This makes latent tree models useful in phylogenetic analysis [37, 74]. In this context, the data typically consist of aligned DNA sequences of length , where each site in the sequence is treated as an independent realisation of the vector . In this case all state spaces are of the form or binary. These applications are not restricted to discrete data. The early evolutionary trees were all built based on morphological characters such as body size. Moreover, with the burst of new genomic data, such as gene expression, phylogenetic models for continuous traits are again becoming important; see  and references therein. Latent tree models in this evolutionary context are also popular in linguistics [68, 76], where a tree represents evolution of languages with modern languages represented by the leaves. The data here typically consists of the acoustic structure of spoken words and are continuous although there are also approaches using syntatic (discrete) data; see, for example, [41, 78]. A related application of latent tree models is in network tomography, where it is used to determine the structure of the connections in the Internet [14, 34]. In this application messages are transmitted by sending packets of bits from a source vertex to different destinations and the correlation in arrival times is used in order to infer the underlying network structure. A common assumption is that the underlying network has a tree structure. Then Gaussian latent tree models form a natural correlation model for arrival times because the correlations diminish with the distance on a tree; see Section 2.1. Typically in this context a special submodel called the Brownian motion tree model is used.

Third, a tree can represent hierarchical structure in complex data sets. This viewpoint was behind the definition of latent tree models that emerged in the machine learning community

[11, 54, 96]. In a rooted tree every internal vertex represents a cluster given by all the leaves that descent from it. We refer to  for an overview of potential applications. We emphasise that in those applications, it is often not realistic to assume that the true data-generating distribution lies in the latent tree model, which leads to some subtleties in inference; see also Section 3.3. Another application in this vein is in computer vision and image processing; see  and reference therein. One promising application along these lines is the use of context in computer vision; see, for example, .

Latent tree models can be also used as a generalization of hidden Markov models. A hidden Markov model (HMM) is a latent tree model on the caterpillar tree in Figure 2(a). Typically hidden Markov models are homogeneous in the sense that the conditional distributions and do not depend on ; see 

. The unobserved part in HMMs follows a Markov chain and so respects a very simple dependency structure. This may be a good approximation of the real dependency structure in the context of time series but is often too restrictive in other applications. Hidden Markov Tree models (HMTMs)

 relax this restriction allowing for any tree structure on the unobserved vector. Still however, like for HMMs, we assume that each internal vertex has a leaf as a child; see Figure 2(b). Models of this type were proposed for wavelet-based statistical signal processing but other applications emerged recently: image processing [20, 71], biomedicine [57, 66], computational lynguistics . In these settings the unobserved vector typically is binary and the observed part is Gaussian.

This leads to another important reason to study latent tree models: many popular models are submodels of the latent tree models. Examples are given by HMMs, HMTMs, all phylogenetic tree models, but also one factor analysis model, latent class models, and Brownian motion tree models. Models of these types are used virtually everywhere: in biostatistics, machine learning, and social sciences. As noted by Wainwright and Jordan in , a more general viewpoint gives a unifying framework for existing algorithms used for these different model classes.

### 1.3 Parsimonious latent tree models

In this section we briefly discuss certain redundancy in discrete latent tree models. We start by discussing the latent class model, which is a special latent tree model where the state space is finite and the underlying tree is a star as in Figure 1(a). In this case the tree is typically rooted at the internal vertex and the parameter consists of the root distribution together with the conditional distributions of the leaves given the internal vertex. The number of latent classes corresponds to the number of states of the unobserved variable . This relatively simple class of models gives us first insights into general latent tree models. The following result shows that, if the number of unobserved classes is high enough the model becomes saturated, that is, it contains all probability distributions over .

###### Proposition 1.

If a given latent class model is not saturated then .

The proof of this result can be recovered from the proof of [96, Theorem 3].

A latent tree model is parsimonious if there is no other latent tree model with a smaller number of parameters that gives the same family of probability distributions over the observed variables. A discrete latent class model is not parsimonious if because, by Proposition 1, every such model is saturated, and the only parsimonious latent tree model over that is saturated is a tree with a single vertex representing the whole vector as a single variable with states.

A discrete latent tree model is regular if the inequality in Proposition 1 holds for any unobserved vertex with neighbours , that is, when

 |Yv|<∏u∈Nv|Yu|/maxu|Yu|. (3)

The argument for the latent class model can be generalised to conclude the following; see :

###### Proposition 2.

Any parsimonious discrete latent tree model is regular.

Proposition 2 substantially reduces the space of possible discrete latent tree models to consider. As demonstrated by  the size of the space of regular models is bounded by . There are two problems with that in practice. First, this space is still very big with no clear structure. Second, no necessary and sufficient conditions to assure parsimony are known in general and so we do not know how good this reduction is. This leads us to a more tractable subclass of discrete latent tree models called general Markov models, which we discuss in the next section.

### 1.4 Gaussian and general Markov models

A general Markov model is a latent tree model for which all are equal and . Models of this type, for or appeared in phylogenetics half a century ago , they were formulated in the most general form over 30 years ago , but only recently they are becoming increasingly popular; see for example [3, 49]. In statistics, general Markov models appeared in the context of causal inference , or simply as the simplest interesting family of graphical models with unobserved variables . As we present here, general Markov models stand out as the tractable class of discrete latent tree models, as much as Gaussian models stand out as the tractable class in the continuous case.

In the previous section we argued that the space of parsimonious latent tree models is not easy to handle. On the other hand, for general Markov models, inequality (3) is satisfied as long as all internal vertices have degree at least three. For general Markov models and Gaussian latent tree models the necessary and sufficient condition for model parsimony is that:

• each unobserved variable has degree at least three,

• any two neighbouring variables are neither functionally related nor independent.

It is standard to assume that the underlying tree has no degree two vertices. In fact, for general Markov models and Gaussian latent tree models we can always suppress degree two vertices without changing the model, and so (A1) is always satisfied; see [100, Section 5.3.4]. In particular, the model defined over a binary rooted tree is equal to the model over the corresponding undirected trivalent tree obtained by suppressing the root. We also always assume the following:

• all are nondegenerate meaning that the distribution of has the full support .

Working without assuming (A2) is sometimes convenient because of the following result, which allows us to focus on learning latent tree models over trivalent trees.

###### Theorem 3.

Every discrete latent tree model satisfying (A1) and (A2) is a submodel of a latent tree model over a trivalent tree that satisfies only (A1). The same applies to Gaussian latent tree models.

A formal proof can be based on the following two observations; see [100, Section 5.2.2] for more details. First, a tree with no degree two vertices can be obtained from a trivalent tree by edge contraction. Second, if in a tree model, there exists a simpler model on a tree obtained from by contracting the edge . This means that every latent tree model can be realised as a submodel over a trivalent tree with some of the vertices identified. For example, consider the discrete latent tree model for the tree on the left of Figure 3. Here one of the internal vertices, labeled with , represents an observed random variable. We can alternatively consider a discrete latent tree model on the right of Figure 3. Here the double edge represent equality between adjacent random variables, and so, the corresponding conditional distribution is degenerate. Directly from the way these models are parameterised, we see that both models are equivalent.

There are numerous advantages of general Markov models. The link to statistical physics, phylogenetics, and Markov processes allows to develop efficient algorithms with strong theoretical guarantees. General Markov models retain very good performance on real-world data . The space of parsimonious general Markov models is also much more tractable than the space of all regular latent tree models over . In some applications we cannot assume that the state spaces of observed variables are equal but we can assume that the unobserved variables have the same state spaces. This holds for HMMs and HMTMs. Many techniques discussed later can be generalised to that case.

For any edge denote by the stochastic matrix representing the conditional distribution . In analogy to Markov chains, is called a transition matrix and each of its rows represents a conditional distribution of given a particular value of . In the general Markov model, each is an arbitrary stochastic matrix. Models used in phylogenetics are usually more constrained. They are generated through a continuous time Markov process on with a given rate matrix , that is a real matrix with row elements summing to zero and all off-diagonal elements nonnegative. In this setting the transition matrices are given by

 Muv=exp(tuvQ), (4)

where is an edge parameter and is the matrix exponential function. The rate matrix typically has some further symmetries. For example, in the Jukes-Cantor model , all off-diagonal entries of are equal; see  for a review of most popular phylogenetic models.

## 2 Second-order moment structure

Many learning algorithms for latent tree models use the second-order structure of the observed distribution, that is, correlations between the observed variables, mutual information, or other aspects of the pairwise marginal distributions. These algorithms typically exploit links to tree metrics. In this section we start by explaining this link and how it can be used to design robust learning algorithms.

### 2.1 Gaussian latent tree model

The earliest example of Gaussian latent tree models is the factor analysis model with a single unobserved factor [8, 88]. More general Gaussian tree models were not studied until more recently [22, 77]. In the fully observed Gaussian tree model the inverse covariance matrix of is very sparse. However, the inverse covariance matrix of the observed subvector has no zeros and it is not convenient to work with. On the other hand, the covariance matrix of provides a great insight into the structure of Gaussian latent tree models and latent tree models in general.

Consider any two vertices of . Using the Markov properties implied by the tree model, the correlation can be written as the product (see, e.g., )

 ρij=∏(u,v)∈¯¯¯¯ijρuv, (5)

where denotes the unique path between and in . Restricting (5) only to pairs of leaves , gives the parameterization of the correlations of the Gaussian latent tree model. In particular, vertices that are far from each other in the tree tend to be less correlated. This appealing property makes this model useful for hierarchical data clustering and network tomography problems as described in Section 1.2.

The constraints on the correlations induced by (5

) are the only nontrivial constraints on Gaussian latent tree models. The variances of the observed variables can be arbitrary and together with the

edge correlations, for all edges , they provide a set of parameters for the latent tree model. The variances of the unobserved variables do not affect the observed distribution and so typically are set to . The model is parsimonious as long as it satisfies (A1) and (A2). The condition (A2) is satisfied if the edge correlations satisfy .

###### Example 1.

Consider the quartet tree in Figure 1(b) and denote the internal vertices by , , where is closer to and . If correlations between the four leaves come from a latent tree model on this quartet tree, then there is a collection of edge correlations , , , , and all with values in such that , , , , , and . In particular

 ρ13ρ24=ρ14ρ23=ρ12ρ34ρ2uv. (6)

The first equality is an example of a nontrivial relation between the observed correlations. The second equality can be used to recover the value of the edge parameter given only the observed correlations, or to establish the inequality .

A systematic study of polynomial constraints defining statistical models is part of algebraic statistics . For Gaussian latent tree models these constraints were studied in [85, Section 6] and .

Correlation matrices of latent Gaussian tree model are closely linked to tree metrics. Given a tree with leaves assign to each of each edges a nonnegative length . With this choice we can now compute the distance between any two leaves summing the lengths of the edges on the unique path between and , that is,

 dij=∑(u,v)∈¯¯¯¯ijduv. (7)

Consider an symmetric matrix with zeros on the diagonal. Call a tree metric if its entries satisfy (7) for some tree and edge lengths.

Let be a correlation matrix in a Gaussian latent tree model. Assume first that for all . Consider a symmetric matrix with . Then (5) translates into (7). Since for all also , and so is a tree metric. If contains zero entries (5) implies that zeros cannot appear arbitrarily. For every three indices if and then also . It follows that the correlation matrix in a Gaussian latent tree model has a block diagonal structure and within each block all entries are non-zero. Each block can be transformed to a tree metric on the corresponding subtree.

The connection to tree metrics will be exploited in many ways. For example, it is a standard result in phylogenetics  that for any tree metric the underlying tree and positive edge lengths can be recovered uniquely. Because , we obtain the following corollary.

###### Theorem 4.

If is a correlation matrix in a Gaussian latent tree model with non-zero entries, then the underlying tree and edge correlations are identified uniquely (up to sign).

To have a concrete example of how it works consider again the quartet tree model in Example 1. Suppose that we are given a correlation matrix from the Gaussian latent tree model but we do not know the underlying tree. By Theorem 3 we can first constrain ourselves to trivalent trees, that is, the three possible quartets: , , and , where this notation indicates how leaves group together. In the first case, by (6), we have

 |ρ12ρ34|≥|ρ13ρ24|=|ρ14ρ23| (8)

and by symmetry we obtain similar relations for the other two trees. Therefore, we can find the underlying tree by computing quantities of the form and choosing the largest. If they all happen to be equal, the underlying tree is the star tree.

We can extend Theorem 4 to arbitrary correlation matrices but here identifiability of the edge correlations is more involved; see  for details. For instance, consider again the model in Example 1. Suppose that for all and and . Then we can easily identify the underlying tree . Indeed, for no other tree with four leaves there is a choice of edge correlations giving precisely this pattern of zeros. Identifying the parameters is harder. We check that must be zero. The other edge correlations must be nonzero but they are identified only up to the relations and .

### 2.2 General Markov models

Tree metrics appear also in the description of discrete latent tree models [9, 52, 56, 82]. For any edge denote by the transition matrix representing the conditional distribution . Denote by the matrix of the marginal distribution of , and by a diagonal matrix with the marginal distribution of on the diagonal. For any two vertices let

 τuv:=det(Puv)√det(PuuPvv), (9)

where the denominator is non-zero by assumption (A3). By essentially the same argument as in [74, Theorem 8.4.3] we obtain the following path-product formula

 τij=∏(u,v)∈¯¯¯¯ijτuvfor all % i,j∈V. (10)

In the case of binary variables,

, and so is the correlation, which implies that (10) reduces to (5).

In general the interpretation of the edge parameters is more complicated. Using the identity we can write

 τuv=detMuv√det(Puu)det(Pvv). (11)

Therefore, we have if and only if . If this is equivalent to independence of and but in general it is a strictly weaker condition. For example, if , and the first two rows of are equal but not equal to the third one, then but and are not independent. Exactly like in the Gaussian case the edge parameters satisfy and the border values consider to functional dependence of and

. Indeed, by the Bayes’ theorem, we have

 Mvu=(Pvv)−1(Muv)TPuu.

With this notation, (11) gives

 τ2uv=det(MuvMuv)det(Puu(Pvv)−1)=detMuvdetMvu.

Because both and

are stochastic matrices, all their eigenvalues lie in the unit circle. In particular,

and it is equal to precisely when is a permutation matrix, or in other words, if and are functionally related.

This again gives a direct link to tree metrics and the following result; see [17, 84].

###### Theorem 5.

For any distribution in a general Markov model the underlying tree can be uniquely recovered given only its 2-way margins.

We implicitly assume that are nonzero but the theorem can be extended to forests.

### 2.3 Linear models

The fact that in the Gaussian and in the binary case the correlation between observed variables decompose as in (5) can be proved using the following fact: if are binary (or jointly Gaussian) then the conditional expectation is a linear function of . As we see in this section the discrete case can be also interpreted in this way, which gives a unifying framework to understand (5) and (10). Moreover, it gives a much larger families of potential models to consider that also admit a path-product formula linking it to tree metrics.

Let each variable in the system be modeled as a random vector in for a fixed . A ternary variable, for example, will take values , , in instead of the typical in . Each variable can be either discrete or continuous but we add a minor requirement that the matrix is positive definite. A linear latent tree model is a latent tree model in which for every edge in the tree, the conditional expectation is an affine function of

. Linear latent tree models include (multivariate) Gaussian latent tree models, Kalman filters, Gaussian mixtures, Poisson mixtures and general Markov models. Models of this type were first discussed in

. Here we propose a slightly different exposition using ideas from [100, Section 4.3].

We define the normalized version of as . Denoting we obtain

 E[¯Yu|Yw]=Σ−1/2uuΣuwΣ−1/2ww¯Yw. (12)

Define . By the law of total expectation, it follows from (12) that . Let be three random variables with values in such that . Then

 E[¯Yu¯YTw]=E[E[¯Yu|Yw](E[¯YTw|Yw])T]=Σ−1/2uuΣuwΣ−1wwΣwvΣ−1/2vv,

which implies that . Applying this argument recursively we conclude that the path-product decomposition of given in (10) holds for any linear latent tree model.

This clearly generalizes the Gaussian case. To see that this also generalises (10

), for each each discrete random variable with

states take and set the state-space to be , where is the origin and are the elements of the standard basis of .

###### Proposition 6.

With the above convention and .

The proposition implies that as defined in this section is equal to as defined in Section 2.2.

###### Proof of the proposition.

Consider the matrix obtained from by elementary row and column operations: add all rows to the first row and all columns to the first column. Basic linear algebra implies . Matrix has the following block structure. The top-left -block is equal to . The bottom-right -block is equal to and the remaining two blocks are and . The formula for the determinant of a block matrix implies that . The proof of the other equality is analogous. ∎

### 2.4 Distance based methods

The maximum likelihood tree topology recovery is NP hard . This has motivated a number of investigations of other tractable methods for learning trees as well as theoretical guarantees on performance. The link between tree metrics and latent tree models described in the previous sections makes it possible to come up with consistent methods to learn a tree that work in polynomial time. This approach dates back to ; see  and references therein.

For a concrete example consider a sample from the Gaussian latent tree model. Given the sample correlation matrix with elements we compute distances . Now use any of the methods to learn a tree metric from observed distances. This gives a tree and edge distances , or equivalently absolute values of the edge correlations . Such a method will be (statistically) consistent given the original tree distance method is consistent, which in this context means that the method outputs the correct tree given a tree metric.

There are many methods that try to recover the underlying tree from noisy distances. The most popular are the Neighbour-Joining (NJ) algorithm and the least-squares method but many other algorithms are available; see Section 7.3 in  for an overview. All popular methods are both well studied and widely implemented, for example, in R; see Section 5.1 in . Most of the methods, including NJ and the least squares, are consistent. We also note in passing that these methods output an undirected tree but rooting is also possible by finding an appropriate outgroup; see  [30, Section 7.3].

An appealing property of this method, as applied in the Gaussian case, is that there is a one-to-one correspondence between edge lengths and model parameters given by edge correlations (up to sign). In the discrete case the situation is more complicated. Here we obtain noisy distances by defining like in (9) with , , and replaced by their sample versions. This gives . Using the NJ algorithm we obtain an estimate of the underlying tree and parameters . However, this is in general not enough to recover all model parameters. A special case when it is possible is so called symmetric discrete distributions. In these submodels the matrix has all off-diagonal entries equal and the root distribution is assumed to be uniform. In statistical mechanics this corresponds to the Potts model, which in the binary case gives the Ising model; see Example 3.2 in .

Although using the NJ algorithm as a tree learning subroutine results in a computationally efficient and consistent method, consistency is only a minor requirement, and other tree learning procedures can be preferred in order to allow for a more sophisticated statistical analysis. An early example is the Dyadic Closure Tree Construction method (DCTC)  and witness-antiwitness method (WAM)  both focusing on learning the underlying tree by learning certain quartets that hold; see the latter paper for more details and references. In  the authors propose two other algorithms: recursive grouping (RG) and CLGrouping. Recursive grouping builds the latent tree recursively by identifying sibling groups using distances . CLGrouping starts with a pre-processing procedure in which a tree over the observed variables is constructed, or more precisely, the Chow–Liu tree. This global step groups the observed vertices that are likely to be close to each other in the true latent tree, thereby guiding subsequent recursive grouping (or equivalent procedures such as neighbour-joining) on much smaller subsets of variables. This results in more accurate and efficient learning of latent trees. This can be further improved by using distance information to learn locally small trees and then glue them together. An example of such a divide-and-conquer algorithm is given in .

The distance based methods to learn the underlying tree are based predominantly on second order margins and so typically are very robust with respect to the sampling error and model misspecifications. In the next section we present other methods to estimate model parameters that use much more information about the underlying distribution.

## 3 Selected theoretical results

Latent tree models, like all models with unobserved variables, suffer from various problems and learning is generally complicated. The complex geometry of models with unobserved variables usually leads to difficulties in establishing the identifiability of their parameters, and the likelihood function has many local maxima, which lie on the boundary of the parameter space; see, for example, [23, 101]. In consequence, standard inference and model selection procedures are not fully justified in this setting. In this section we discuss parameter identifiability, sample complexity, and model selection methods. These three seemingly unrelated topics all deliver one important message: the class of latent tree models is well behaved from the statistical point of view as long as all the edge correlations are sufficiently large in the absolute value. Otherwise, it should be used with caution. Some further issues with the likelihood function for this model class will be discussed in Section 4.1.

### 3.1 Identifiability

is identifiable if implies . In other words, the parameterization map is a bijection between the parameter space and the model. Latent tree models are never identifiable. This can be seen for latent class models, where permuting labels of the unobserved variable makes no difference in the observed distribution; see, for example, . This is known as the label swapping problem. The label swapping is not a serious problem in practice. We can always take account of it by restricting the parameter space. However, this still does not make the model identifiable because there are special subspaces in the parameter space that map to the same observed distribution. We illustrate this issue in the simplest possible example.

###### Example 2.

Consider the Gaussian latent tree model on the star tree with three leaves, see Figure 1(a). Denote and for . By (5), this model is parameterized by , , and . If the observed correlations are non-zero we can identify up to the sign and then the other parameters as follows

 ρ21=ρ12ρ13ρ23,ρ2=ρ12ρ1,ρ3=ρ13ρ1.

Suppose now that some observed correlations vanish. The form of the parameterization implies that it is impossible for only one of them to vanish. If two correlations are zero, say , then the set of all triples mapping to is a smooth one-dimensional subset given by and . Suppose now that all observed correlations are zero. Then the corresponding parameters form the union of three intervals

 {ρ1=ρ2=0}∪{ρ1=ρ3=0}∪{ρ2=ρ3=0}.

This example motivates the following definition.

###### Definition 7.

A model is generically identifiable, if the parameterization map is finite-to-one everywhere outside of a measure zero set.

Geometrically, generic identifiability means that for a typical point in the model, its preimage under the parameterization map, also called a fiber, is a finite collection of points. Showing that the Gaussian latent tree model satisfying (A1) is generically identifiable can be done by arguments as in Example 2

. For general Markov models generic identifiability is more subtle and, in general, the second-order moments contain not enough information about the underlying parameters. It turns out that three-way margins are already enough. The following is the main result of

.

###### Theorem 8.

Every general Markov model satisfying (A1) is generically identifiable. In fact, to identify the parameters it is enough to know the -way marginal distributions.

Theorem 8 does not cover the case when the state-spaces are allowed to vary. In this case techniques of  may be useful to establish identifiability.

The analysis of when exactly identifiability fails can be in general complicated. In the Gaussian and the binary case these special points correspond to some correlations being zero . As illustrated by Example 2, depending on the situation the corresponding fiber can be either a smooth subset of the parameter space or it can be singular. These theoretical results and examples provide the following insight. If the true data-generating distribution is characterised by high correlations between variables, it is also far from any of the special singular points. However, if some variables have a low degree of dependence, then estimation may become difficult and standard asymptotic theory breaks down.

### 3.2 Guarantees for tree reconstruction

A basic question regarding learning of latent tree models is that of the sample complexity of the tree reconstruction problem. Given an estimator of the true tree we want to assure that for some fixed small . This is only possible if the sample size is big enough. It is known that irrespective of the method is necessary [32, 58] but it is typically not enough. The first systematic study of when the logarithmic bound is sufficient was offered in  for the binary symmetric model and in  for general Markov models. The link to tree metrics shows that, in order to recover the underlying tree with high probability, the edge lengths cannot be too small or too large. For linear latent tree models this means that the edge parameters must satisfy for some . Sample bounds we discuss will always necessarily depend on the number of observed variables and the parameters .

One of the important concepts developed in [32, 33] is that of the tree depth: the depth of a tree with vertices and leaves is For example, the depth of the tree in Figure 1(c) is and of both trees in Figure 2 is . Latent tree models for trees with few unobserved variables or small depth are generally easier to learn. This leads to one group of results assuming that the depth of the underlying tree is , that is, constant in the number of leaves of . It was shown in [33, Theorem 14] that, under constant tree depth, there is an algorithm for general Markov models whose sample complexity is logarithmic in . This has been generalised to the Gaussian case in .

Similarly strong results are possible without assuming constant depth. It was first conjectured by Mike Steel  that for the binary symmetric latent tree model the sample complexity is logarithmic in as long as for some , that is, when the parameters lie in the Kesten–Stigum (KS) regime. This conjecture was proved in [26, 59]. In the subsequent work, it has been shown in multiple scenarios, including the Gaussian case, that in the KS regime, high probability tree topology reconstruction may be achieved with samples; see  and references therein. The KS regime plays also an important role in the sample complexity analysis of the maximum likelihood estimator. In general the sample complexity is polynomial in under very general assumptions. However, for symmetric models and under a discretisation assumption, it becomes logarithmic in if the true distribution lies in the KS regime .

### 3.3 Model selection

As presented in the previous section, a lot of effort in the literature is put to obtain performance guarantees for the proposed learning algorithms. This analysis is done under assumption that the data come from a latent tree model. Although this assumption seems to be reasonable for phylogenetics and several other applications mentioned here, it is certainly not so for many other kinds of data (e.g. survey data from medicine and social sciences) for which these models are used, c.f. Section 1.2. In all these cases model selection techniques tend to outperform algorithms that aim at finding the “true” tree .

An example of a model selection algorithm is EAST (Expansion, Adjustment and Simplification until Termination) , which aims to find the model with the highest Bayesian Information Criterion (BIC) score. Given a random sample the BIC is

 BIC(M(T,Y))=ℓ(^θ,T)−d2logn,

where is the maximum likelihood estimator, and is the dimension of the model . Computing the dimension of a model with unobserved variables is not always easy because it need not correspond to the number of parameters in the model. In that case a more detailed study of the generic rank of the Jacobian of the parameterization is needed. However, for general Markov models, Theorem 8 can be used to show that simple parameter count does suffice to compute the dimension of the model whenever (A1) holds. For Gaussian latent tree models the analysis in Section 2.1 showed that the dimension is , where is the set of edges of the underlying tree and is the number of its leaves.

Theoretical importance of the BIC criterion  is that it provides an asymptotic approximation to the marginal likelihood of the model, and so it can guide model selection in the Bayesian setting. However, latent tree models lead to a new difficulty in that the Fisher information matrix of a latent tree model is singular along certain submodels. Such singularities invalidate the mathematical arguments that lead to the Bayesian information criterion; see, for example, [28, 55, 92]. This again shows that BIC must be used with caution in the case of week correlations. In that case the BIC may be too conservative and select too small models. Correcting BIC to account for singularities involves an indepth study of the model geometry. For binary latent tree models this has been done in  and for Gaussian latent tree models in . In general, a simpler adjustment can be done by replacing BIC with Watanabe’s WAIC .

## 4 Estimation and inference

In Section 2.4 we described some natural tree-metric based approaches to learn latent tree models. In this section we briefly describe other popular learning algorithms.

### 4.1 Fixed tree structure

The maximum likelihood estimator is one of the most popular estimators of the continuous parameter. It cannot be computed in a closed form and the likelihood function is very complicated to analyse directly; see, for example, [23, 101] . However, there are various numerical methods to maximise the likelihood function that were developed in the context of latent tree models or, more generally, Bayesian networks with unobserved variables. The most natural choice is the EM algorithm, which is easy to set up; see, for example, Section 3.1.1 in . There are also several methods building upon the idea of the EM algorithm. For example the progressive EM algorithm of  uses a method of moments estimator as a subroutine, which leads to a more computationally efficient estimator.

As with all other EM-based methods, these approaches depend on the initialisation and suffer from the possibility of being trapped in local optima. This algorithm seems to work well in the case when the observed variables are highly correlated . The situation becomes much more complicated in the presence of weak correlations as the numerical procedures become unstable. Another problem is that the maximum likelihood estimator often lies on the boundary of the parameter space , which has many important consequences. First, the gradient of the likelihood function does not vanish at such a point and so the standard asymptotics does not apply. Second, there may be relatively distant points in the parameter space that give a similar value of the likelihood function as the maximum likelihood estimator. Finally, the boundary points typically correspond to situations when the distribution of the unobserved vector is degenerate. In many applications this may be problematic as a natural interpretation for the unobserved process may be lost.

### 4.2 The Structural EM algorithm

Suppose that given a random sample of size , we are interested in finding the tree that maximizes the likelihood function over all fully-observed tree models. For every tree the maximum likelihood estimate is easily obtained; see [53, Section 4.4.2 and Section 5.2.1]. In the naive approach we can search over all possible trees with a given number of vertices and find the one that gives the highest value of the likelihood function. The Chow–Liu algorithm  is a remarkably simple algorithm, which gives an efficient way to find the best tree approximation that maximizes the likelihood.

The original approach was proposed for discrete data but it can be easily extended to the Gaussian case . It only uses the fact that the MLE decomposes according to a tree so it can be also used in other similar scenarios to find best BIC and AIC trees ; see  for further references and implementations. This approach boils down to using the Kruskal’s algorithm to find the maximum cost spanning tree of a complete graph with edge weights given by sample mutual informations :

• Compute all mutual informations of the sample distribution and order them from the largest to the smallest.

• Move along this ordered sequence adding subsequently the corresponding edges unless adding an edge introduces a cycle. Stop when no more edges can be added.

If all mutual informations are different, then there is a unique best solution. If some of the weights are equal, then multiple solutions are possible, but they will all give the same value of the likelihood function. In the Gaussian case the mutual information is a simple monotone function of the corresponding correlations squared. Therefore, the same tree will be obtained after replacing mutual informations in step 1 above with squares of sample correlations.

A natural idea is to use the Chow–Liu algorithm to learn latent tree models by using an EM-type algorithm. The structural EM algorithm  is a numerical procedure to maximize the likelihood simultaneously over the continuous parameter and the discrete parameter . It starts from a given parameter value and moves at each step strictly increasing the likelihood unless it is already in a local optimum. The E-step of the algorithm is the standard E-step of the EM algorithm. The M-step follows essentially the Chow–Liu algorithm with several modifications. A minor problem with the Chow–Liu algorithm, when used in the EM algorithm, is that it does not get any information about which vertices represent unobserved variables and so it often outputs a tree whose leaves can be potentially unobserved or internal vertices that can represent observed random variables. In this case it is easy to provide a tree, with leaves precisely corresponding to observed vertices, that gives the same observed likelihood. This procedure is described in more detail for the discrete case in [39, Section 5]. We describe the general idea in the Gaussian case. Suppose that the Chow–Liu algorithm outputs a tree with edge correlations , then:

• Remove all degree one vertices that represent unobserved vertices.

• If there is an induced chain , where stands for any vertex of degree at least three, then replace this chain with a single edge between and . Set the correlation equal to the product .

• If there is an internal vertex representing an observed random variable then add an auxiliary copy of and an edge . Set .

This operation gives a leaf-labeled tree that leads to the same observed likelihood as and we take it as the output of the M-step. If we are interested only in trivalent trees, the above procedure can be easily modified so that the output is a trivalent tree; c.f. Theorem 3.

### 4.3 Phylogenetic invariants

Given latent tree model over , we associate to it the set of all polynomials vanishing on . The polynomials are expressed in terms of correlations in the Gaussian case and in terms of the raw probabilities in the discrete case. Every polynomial is called a phylogenetic invariant. Equation (6) provides an example of an equation that must hold for a Gaussian latent tree model over a quartet tree. It is a basic result in algebraic geometry that is finitely generated, that is, it admits a finite basis of polynomials such that every polynomial in is a polynomial combination of the .

The basic idea behind application of phylogenetic invariants is as follows. Given independent observations of we compute the sample distribution

, which, by the law of large numbers, converges almost surely to the true data generating distribution

. Because for every , for large also . The methods proposed in the phylogenetic literature are mainly simple diagnostic tests that work with a given fixed finite set of invariants in , which do not necessarily generate . There is now considerable literature on the method of phylogenetic invariants and for many models all defining polynomials are understood. We refer to [2, 86] for an overview.

The advantage of this approach is that the method based on phylogenetic invariants does not require parameter estimation. The disadvantage is that to large extent statistical theory behind their use is lacking. The method of phylogenetic invariants gives a way to select the best tree under a given criterion. It does not, however, give a way of quantifying how well the chosen tree fits the data because, in general, the distribution of phylogenetic invariants is too hard to analyze.

Recently there has been some effort aimed at organizing the statistical theory behind phylogenetic invariants. For example, it has been observed by many authors that not all invariants are equally important and from the statistical point of view there is no sense in working with a full generating set of . For example, invariants linking directly to tree metrics, called the edge invariants, tend to be more robust with respect to the sampling error and are enough to distinguish between different models . To test the edge invariants, 

uses the singular value decomposition and the Frobenius norm to compute the distance of a matrix to the set of matrices of certain rank. Recently this method has been further improved by

. In its current form the method is robust and simulations show that it outperforms most of the commonly used methods.

Focusing only on quadratic invariants allows us to generalise directly asymptotic chi-square tests for independence in a contingency table; see

. In the Gaussian setting we can proceed as follows. For any four leaves of , forms a quartet in if the paths and have no vertices in common. It is clear from (5) and (10) that for any quartet we have that in the Gaussian case and in the discrete case. For example (6) gives the equation that holds for the quartet tree in Example 1. Equations of this form are called tetrads . In the context of latent tree models using tetrads was suggested already by Judea Pearl [64, Section 8.3.5] but with no statistical guidance. An example of a statistically guided quartet-based analysis was given recently for Gaussian latent tree models .

Finally, the geometric description of latent tree models involves not only polynomial equalities but also polynomial inequalities. These inequalities cut out a large portion of the space described solely by equalities and therefore they should not be neglected . The inequalities are harder to study than equalities but they are also understood well for general Markov models [5, 75, 101]. Recently  showed how basic inequalities can be easily tested within the Bayesian framework to obtain a preliminary assessment of whether the data come from a Gaussian latent tree model.

## 5 Discussion

We gave a concise overview of the theory of latent tree models. We argued that, from the theoretical and the practical point of view, the general Markov models, and more generally, linear latent tree models, form the most important subfamily. We showed that for linear latent tree models the link to tree metrics provides a wide variety of learning procedures. The geometric viewpoint gives important insights but was not covered here in much detail. We refer to  for further details. A more algorithmic overview of the latent tree model class is provided in 

. We also skipped other research directions that we believe will become increasingly important in the coming years. Tensor representations of discrete latent tree models help to design learning procedures for latent tree models and more general graphical models with unobserved variables

[7, 47, 48, 63]. In the numerical analysis community a closely related concept of hierarchical tensors has become popular . There are also several recent approaches that introduce the nonparametric setting for latent tree models [80, 81].

## Acknowledgements

I would like to thank Anima Anandkumar, Sebastien Roch, and Nevin L. Zhang for helpful comments, clarifications, and literature suggestions. Comments from the anonymous referees allowed to substantially improve the manuscript.

## References

•  Elizabeth S. Allman, Catherine Matias, and John A. Rhodes. Identifiability of parameters in latent structure models with many observed variables. Ann. Statist., 37(6A):3099–3132, 2009.
•  Elizabeth S. Allman and John A. Rhodes. Phylogenetic invariants. In Reconstructing Evolution: New Mathematical and Computational Advances, pages 108–146. Oxford Univ. Press, Oxford, 2007.
•  Elizabeth S. Allman and John A. Rhodes. Phylogenetic ideals and varieties for the general Markov model. Adv. in Appl. Math., 40(2):127–148, 2008.
•  Elizabeth S. Allman, John A. Rhodes, Bernd Sturmfels, and Piotr Zwiernik. Tensors of nonnegative rank two. Linear Algebra Appl., 473:37–53, 2015.
•  Elizabeth S. Allman, John A. Rhodes, and Amelia Taylor. A semialgebraic description of the general Markov model on phylogenetic trees. SIAM J. Discrete Math., 28(2):736–755, 2014.
•  Animashree Anandkumar, Kamalika Chaudhuri, Daniel J Hsu, Sham M Kakade, Le Song, and Tong Zhang. Spectral methods for learning multivariate latent tree structure. In Advances in Neural Information Processing Systems, pages 2025–2033, 2011.
•  Animashree Anandkumar, Rong Ge, Daniel Hsu, Sham M. Kakade, and Matus Telgarsky. Tensor decompositions for learning latent variable models. J. Mach. Learn. Res., 15:2773–2832, 2014.
•  T. W. Anderson and Herman Rubin. Statistical inference in factor analysis. In Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, 1954–1955, vol. V, pages 111–150. University of California Press, Berkeley and Los Angeles, 1956.
•  Daniel Barry and J. A. Hartigan. Statistical analysis of hominoid molecular evolution. Statist. Sci., 2(2):191–210, 1987. With comments by Stephen Portnoy and Joseph Felsenstein and a reply by the authors.
•  Christopher M Bishop. Pattern recognition and machine learning. Springer, 2006.
•  Christopher M Bishop and Michael E Tipping.

A hierarchical latent variable model for data visualization.

IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(3):281–293, 1998.
•  Peter Buneman. The recovery of trees from measures of dissimilarity. In F. Hodson et al., editor, Mathematics in the Archaeological and Historical Sciences, pages 387–395. Edinburgh University Press, 1971.
•  Marta Casanellas and Jesús Fernández-Sánchez. Relevant phylogenetic invariants of evolutionary models. J. Math. Pures Appl. (9), 96(3):207–229, 2011.
•  Rui Castro, Mark Coates, Gang Liang, Robert Nowak, and Bin Yu. Network tomography: recent developments. Statistical Science, pages 499–517, 2004.
•  LL Cavalli-Sforza and AWF Edwards. Phylogenetic analysis: models and estimation procedures. Evolution, 21(3):550–570, 1967.
•  Joseph T. Chang. Full reconstruction of Markov models on evolutionary trees: Identifiability and consistency. Mathematical Biosciences, 137(1):51–73, 1996.
•  Joseph T Chang and John A Hartigan. Reconstruction of evolutionary trees from pairwise distributions on current species. In Computing science and statistics: Proceedings of the 23rd symposium on the interface, volume 254, page 257. Interface Foundation, Fairfax Station, VA, 1991.
•  Peixian Chen, Nevin L. Zhang, Leonard Poon, and Zhourong Chen. Progressive EM for latent tree models and hierarchical topic detection. In AAAI, pages 1498–1504, 2016.
•  Tao Chen, Nevin L Zhang, Tengfei Liu, Kin Man Poon, and Yi Wang. Model-based multidimensional clustering of categorical data. Artificial Intelligence, 176(1):2246–2269, 2012.
•  Hyeokho Choi and Richard G. Baraniuk. Multiscale image segmentation using wavelet-domain hidden Markov models. IEEE Trans. Image Process., 10(9):1309–1321, 2001.
•  Myung Jin Choi, Joseph J Lim, Antonio Torralba, and Alan S. Willsky. Exploiting hierarchical context on a large database of object categories. In Computer Vision and Pattern Recognition (CVPR), pages 129–136. IEEE, 2010.
•  Myung Jin Choi, Vincent Y. F. Tan, Animashree Anandkumar, and Alan S. Willsky. Learning latent tree graphical models. J. Mach. Learn. Res., 12:1771–1812, 2011.
•  Benny Chor, Michael D. Hendy, Barbara R. Holland, and David Penny. Multiple maxima of likelihood in phylogenetic trees: An analytic approach. Molecular Biology and Evolution, 17(10):1529–1541, 2000.
•  C. K. Chow and C. N. Liu. Approximating discrete probability distributions with dependence trees. IEEE Trans. Inform. Theory, 14:462–467, 1968.
•  Matthew S. Crouse, Robert D. Nowak, and Richard G. Baraniuk. Wavelet-based statistical signal processing using hidden Markov models. IEEE Trans. Signal Process., 46(4):886–902, 1998.
•  Constantinos Daskalakis, Elchanan Mossel, and Sebastien Roch. Phylogenies without branch bounds: Contracting the short, pruning the deep. SIAM Journal on Discrete Mathematics, 25(2):872–893, 2011.
•  Mathias Drton, Shaowei Lin, Luca Weihs, and Piotr Zwiernik. Marginal likelihood and model selection for Gaussian latent tree and forest models. Bernoulli, 23(2):1202–1232, 2017.
•  Mathias Drton and Martyn Plummer. A Bayesian information criterion for singular models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 79(2):323–380, 2017.
•  Mathias Drton, Bernd Sturmfels, and Seth Sullivant. Lectures on Algebraic Statistics. Oberwolfach Seminars Series. Birkhauser Verlag AG, 2009.
•  Richard Durbin, Anders Krogh, Graeme Mitchison, and Sean R. Eddy. Biological Sequence Analysis: Probabilistic Models of Proteins and Nucleic Acids. Cambridge University Press, 1998.
•  David Edwards, Gabriel CG De Abreu, and Rodrigo Labouriau. Selecting high-dimensional mixed graphical models using minimal aic or bic forests. BMC bioinformatics, 11(1):1, 2010.
•  Péter L. Erdős, Michael A. Steel, László A. Székely, and Tandy J. Warnow. A few logs suffice to build (almost) all trees (part 1). Random Structures and Algorithms, 14(2):153–184, 1999.
•  Péter L. Erdős, Michael A. Steel, László A. Székely, and Tandy J. Warnow. A few logs suffice to build (almost) all trees (part 2). Theoretical Computer Science, 221(1):77 – 118, 1999.
•  Brian Eriksson, Gautam Dasarathy, Paul Barford, and Robert Nowak. Toward the practical use of network tomography for internet topology discovery. In INFOCOM, 2010 Proceedings IEEE, pages 1–9. IEEE, 2010.
•  Nicholas Eriksson. Using Invariants for Phylogenetic Tree Construction, volume 149 of The IMA Volumes in Mathematics and Its Applications, pages 89–108. Springer, 2007.
•  Robin J. Evans. Graphs for margins of Bayesian networks. Scandinavian Journal of Statistics, 43(3):625–648, 2016. 10.1111/sjos.12194.
•  Joseph Felsenstein. Inferring phylogenies. Sinauer Associates Sunderland, 2004.
•  Jesús Fernández-Sánchez and Marta Casanellas. Invariant versus classical approaches when evolution is heterogeneous across sites and lineages. arXiv:1405.6546, 2014.
•  Nir Friedman, Matan Ninio, Itsik Pe’er, and Tal Pupko. A structural EM algorithm for phylogenetic inference. Journal of Computational Biology, 9(2):331–353, 2002.
•  Morten Frydenberg. The chain graph Markov property. Scand. J. Statist., 17(4):333–353, 1990.
•  Cristina Guardiano and Giuseppe Longobardi. Parametric comparison and language taxonomy. In Montserrat Batllori, Maria-Lluïsa Hernanz, Carme Picallo, and Francesc Roca, editors, Grammaticalization and Parametric Variation, pages 149–174. Oxford University Press, Aug 2005.
•  Wolfgang Hackbusch and Stefan Kühn. A new scheme for the tensor representation. Journal of Fourier analysis and applications, 15(5):706–722, 2009.
•  Gordon Hiscott, Colin Fox, Matthew Parry, and David Bryant. Efficient recycled algorithms for quantitative trait models on phylogenies. Genome biology and evolution, 8(5):1338–1350, 2016.
•  Søren Højsgaard, David Edwards, and Steffen Lauritzen. Graphical models with R. Springer Science & Business Media, 2012.
•  Barbara R. Holland, Peter D. Jarvis, and Jeremy G. Sumner. Low-parameter phylogenetic inference under the general Markov model. Systematic Biology, 62(1):78–92, 2013.
•  Furong Huang, U.N. Niranjan, Ioakeim Perros, Robert Chen, Jimeng Sun, and Anima Anandkumar. Scalable latent tree model and its application to health analytics. arXiv:1406.4566, 2014.
•  Mariya Ishteva, Haesun Park, and Le Song. Unfolding latent tree structures using 4th order tensors. In ICML (3), pages 316–324, 2013.
•  Mariya Ishteva, L Song, H Park, A Parikh, and E Xing. Hierarchical tensor decomposition of latent tree graphical models. In The 30th International Conference on Machine Learning (ICML 2013), 2013.
•  Vivek Jayaswal, John Robinson, and Lars S. Jermiin. Estimation of phylogeny and invariant sites under the general Markov model of nucleotide sequence evolution. Systematic Biology, 56(2):155–162, 2007.
•  Thomas H. Jukes and Charles R. Cantor. Evolution of protein molecules. Mammalian Protein Metabolism, 3:21–132, 1969.
•  Joseph B Kruskal. On the shortest spanning subtree of a graph and the traveling salesman problem. Proceedings of the American Mathematical Society, 7(1):48–50, 1956.
•  James A. Lake. Reconstructing evolutionary trees from DNA and protein sequences: Paralinear distances. Proceedings of the National Academy of Sciences of the United States of America, 91(4):1455–1459, 1994.
•  Steffen L. Lauritzen. Graphical Models, volume 17 of Oxford Statistical Science Series. Oxford University Press, 1996. Oxford Science Publications.
•  Neil D. Lawrence.

Gaussian process latent variable models for visualisation of high dimensional data.

Advances in Neural Information Processing Systems, 16(3):329–336, 2004.
•  Shaowei Lin. Ideal-theoretic strategies for asymptotic approximation of marginal likelihood integrals. Journal of Algebraic Statistics, 8(1), 2017.
•  Peter J. Lockhart, Michael A. Steel, Michael D. Hendy, and David Penny. Recovering evolutionary trees under a more realistic model of sequence evolution. Molecular Biology and Evolution, 11(4):605–612, 1994.
•  Mahender K Makhijani, Niranjan Balu, Kiyofumi Yamada, Chun Yuan, and Krishna S Nayak. Accelerated 3d merge carotid imaging using compressed sensing with a hidden Markov tree model. Journal of Magnetic Resonance Imaging, 36(5):1194–1202, 2012.
•  Elchanan Mossel. On the impossibility of reconstructing ancestral data and phylogenies. Journal of Computational Biology, 10(5):669–676, 2003.
•  Elchanan Mossel. Phase transitions in phylogeny. Transactions of the American Mathematical Society, 356(6):2379–2404, 2004.
•  Elchanan Mossel, Sébastien Roch, and Allan Sly. Robust estimation of latent tree graphical models: Inferring hidden states with inexact parameters. IEEE Trans. Inform. Theory, 59(7):4357–4373, 2013.
•  Raphaël Mourad, Christine Sinoquet, Nevin Lianwen Zhang, Tengfei Liu, Philippe Leray, et al. A survey on latent tree models and applications. J. Artif. Intell. Res.(JAIR), 47:157–203, 2013.
•  Emmanuel Paradis. Analysis of Phylogenetics and Evolution with R. Springer Science & Business Media, 2011.
•  Ankur P Parikh, Le Song, and Eric P Xing. A spectral algorithm for latent tree graphical models. In Proceedings of the 28th International Conference on Machine Learning (ICML-11), pages 1065–1072, 2011.
•  Judea Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. The Morgan Kaufmann Series in Representation and Reasoning. Morgan Kaufmann, San Mateo, CA, 1988.
•  Judea Pearl. Causality: Models, Reasoning, and Inference. Cambridge University Press, New York, NY, USA, 2000.
•  Michael Pfeiffer, Marion Betizeau, Julie Waltispurger, Sabina Sara Pfister, Rodney J. Douglas, Henry Kennedy, and Colette Dehay. Unsupervised lineage-based characterization of primate precursors reveals high proliferative and morphological diversity in the OSVZ. Journal of Comparative Neurology, 524(3):535–563, 2016.
•  Lawrence R Rabiner. A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE, 77(2):257–286, 1989.
•  Don Ringe, Tandy Warnow, and Ann Taylor. Indo-european and computational cladistics. Transactions of the Philological Society, 100(1):59–129, 2002.
•  Sebastien Roch. A short proof that phylogenetic tree reconstruction by maximum likelihood is hard. IEEE/ACM Trans. Comput. Biol. Bioinformatics, 3(1):92–, January 2006.
•  Sebastien Roch and Allan Sly. Phase transition in the sample complexity of likelihood-based phylogeny inference. arXiv:1508.01964, 2015.
•  Justin K Romberg, Hyeokho Choi, and Richard G Baraniuk. Bayesian tree-structured image modeling using wavelet-domain hidden Markov models. IEEE Transactions on image processing, 10(7):1056–1068, 2001.
•  David Sankoff. Designer invariants for large phylogenies. Molecular Biology and Evolution, 7(3):255, 1990.
•  Gideon Schwarz. Estimating the dimension of a model. Annals of Statistics, 6(2):461–464, 1978.
•  Charles Semple and Mike Steel. Phylogenetics, volume 24 of Oxford Lecture Series in Mathematics and Its Applications. Oxford University Press, Oxford, 2003.
•  Raffaella Settimi and Jim Q. Smith. Geometry, moments and conditional independence trees with hidden variables. Ann. Statist., 28(4):1179–1205, 2000.
•  Nathaniel Shiers, John AD Aston, Jim Q Smith, and John S Coleman. Gaussian tree constraints applied to acoustic linguistic functional data.

Journal of Multivariate Analysis

, 154:199–215, 2017.
•  Nathaniel Shiers, Piotr Zwiernik, John A. Aston, and James Q. Smith. The correlation space of Gaussian latent tree models and model selection without fitting. Biometrika, 2016.
•  Kevin Shu, Sharjeel Aziz, Vy-Luan Huynh, David Warrick, and Matilde Marcolli. Syntactic phylogenetic trees. arXiv:1607.02791, 2016.
•  Ricardo Silva, Richard Scheine, Clark Glymour, and Peter Spirtes. Learning the structure of linear latent variable models. Journal of Machine Learning Research, 7(Feb):191–246, 2006.
•  Le Song, Animashree Anandkumar, Bo Dai, and Bo Xie. Nonparametric estimation of multi-view latent variable models. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), pages 640–648, 2014.
•  Le Song, Eric P. Xing, and Ankur P. Parikh. Kernel embeddings of latent tree graphical models. In J. Shawe-Taylor, R. S. Zemel, P. L. Bartlett, F. Pereira, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 24, pages 2708–2716. Curran Associates, Inc., 2011.
•  M. Steel. Recovering a tree from the leaf colourations it generates under a Markov model. Applied Mathematics Letters, 7(2):19 – 23, 1994.
•  M Steel. My favourite conjecture. Preprint, 2001.
•  M. Steel, M.D. Hendy, and D. Penny. Invertible models of sequence evolution. Mathematical and Information Science report 93/02, Massey University, 1993.
•  Seth Sullivant. Algebraic geometry of Gaussian Bayesian networks. Advances in Applied Mathematics, 40(4):482–513, 2008.
•  Jeremy G Sumner, Amelia Taylor, Barbara R Holland, and Peter D Jarvis. Developing a statistically powerful measure for quartet tree inference using phylogenetic identities and Markov invariants. Journal of Mathematical Biology, pages 1–36, 2017.
•  Vincent Y. F. Tan, Animashree Anandkumar, and Alan S. Willsky. Learning high-dimensional Markov forest distributions: analysis of error rates. J. Mach. Learn. Res., 12:1617–1653, 2011.
•  Louis L. Thurstone. The vectors of mind. Psychological Review, 41(1):1, 1934.
•  Thomas S. Verma and Judea Pearl. Equivalence and Synthesis of Causal Models. In Piero P. Bonissone, Max Henrion, Laveen N. Kanal, and John F. Lemmer, editors, UAI ’90: Proceedings of the Sixth Annual Conference on Uncertainty in Artificial Intelligence, MIT, Cambridge, MA, USA, July 27-29, 1990. Elsevier, October 1991.
•  Martin J. Wainwright and Michael I. Jordan. Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1-2):1–305, 2008.
•  Yi Wang and Nevin L. Zhang. Severity of local maxima for the EM algorithm: Experiences with hierarchical latent class models. In Probabilistic Graphical Models, pages 301–308. Citeseer, 2006.
•  Sumio Watanabe.

Algebraic Geometry and Statistical Learning Theory

.
Number 25 in Cambridge Monographs on Applied and Computational Mathematics. Cambridge University Press, 2009. ISBN-13: 9780521864671.
•  Sumio Watanabe. A widely applicable Bayesian information criterion. Journal of Machine Learning Research, 14(Mar):867–897, 2013.
•  Alan S. Willsky. Multiresolution Markov models for signal and image processing. Proceedings of the IEEE, 90(8):1396–1458, 2002.
•  Zdeněk Žabokrtskỳ and Martin Popel. Hidden Markov tree model in dependency-based machine translation. In Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, pages 145–148. Association for Computational Linguistics, 2009.
•  Nevin L. Zhang.

Hierarchical latent class models for cluster analysis.

J. Mach. Learn. Res., 5:697–723, 2003/04.
•  Nevin L Zhang and Leonard KM Poon. Latent tree analysis. In AAAI, pages 4891–4898, 2017.
•  Liwen Zou, Edward Susko, Chris Field, and Andrew J. Roger. The parameters of the Barry and Hartigan general Markov model are statistically nonidentifiable. Systematic Biology, 2011.
•  Piotr Zwiernik. Asymptotic behaviour of the marginal likelihood for general Markov models. J. Mach. Learn. Res., 12:3283–3310, 2011.
•  Piotr Zwiernik. Semialgebraic statistics and latent tree models, volume 146 of Monographs on Statistics and Applied Probability. Chapman & Hall/CRC, 2016.
•  Piotr Zwiernik and Jim Q. Smith. Implicit inequality constraints in a binary tree model. Electron. J. Statist., 5:1276–1312, 2011.
•  Piotr Zwiernik and Jim Q. Smith. Tree-cumulants and the geometry of binary tree models. Bernoulli, 18(1):290–321, January 2012.