1 Introduction
Dynamic text networks have been widely studied in recent years, primarily because the Internet stores textual data in a way that allows links between different documents. Articles on the Wikipedia (Hoffman et al., 2010), citation networks in journal articles (Moody, 2004), and linked blog posts (Latouche et al., 2011) are examples of dynamic text networks, or networks of documents that are generated over time. But each application has idiosyncratic features, such as the structure of the links and the nature of the time varying documents, so analysis typically requires bespoke models that directly address those aspects.
This article studies dynamic topic structure and the network properties of the top 467 US political blogs in 2012. Some key features of this data set are (1) topics, such as presidential election news, that evolve over time and (2) community structure among bloggers with similar interests. We develop a bespoke Bayesian model for the dynamic interaction between text and network structure, and examine the dynamics of both the discourse and the community structure among the bloggers.
Our approach combines a topic model and a network model. A topic model infers the unobserved topic assignments of a set of documents (in this case, blog posts) from the text. And a network model infers communities among the nodes (in this case, blogs that tend to link to one another). In combination, we find blocks of blogs that tend to post on the same topics and which link with one another. These blocks, which we call topic interest blocks, allow one to examine sets of similar blogs, such as those that post only on the 2012 election or those that are only interested in both the Middle East and foreign policy. Topic interest blocks allow text content to guide community discovery and link patterns to guide topic learning.
We begin with a review of terminology in topic modeling. A corpus is a collection of documents. A document, in our case a post, is a collection of tokens, which consist of words and ngrams, which are sets of words that commonly appear together (“President of the United States” is a common 5gram). In our application, a blog produces posts. A topic
is a distribution over the tokens in the corpus. Typically, a post concerns a single topic. One such topic might be described as “the 2012 election”, but this labeling is usually done subjectively, after estimation of the topic distributions, based on the highprobability tokens. For example, the 2012 election topic might put high probability on “Gingrich”, “Santorum”, “Cain” and “primaries”
^{1}^{1}1Gingrich, Santorum and Cain all refer to candidates in the 2012 Republican presidential primary..An early and influential topic model is Latent Dirichlet Allocation (LDA), proposed in Blei et al. (2003)
. It is a bagofwords model, since the order of the tokens is ignored. LDA assumes that the tokens in a document are drawn at random from a topic. If a document is about more than one topic, then the tokens are drawn from multiple topics with topic proportions that must be estimated. The LDA generative model can produce a document that is 70% about the 2012 election topic and 30% about a Supreme Court topic by repeatedly tossing a coin with probability 0.7 of coming up heads. When it is heads, LDA draws a word from the 2012 election distribution; otherwise, it draws from the Supreme Court distribution. Markov chain Monte Carlo allows one to reverse the generative model, so that given a corpus of documents, one can estimate the distribution corresponding to each topic, and, for each document, the proportion of that document that is drawn from each topic.
In our application, topic specific word probabilities evolve over time—the token “Gingrich” is more probable early in 2012 than later, when he dropped out. Blei and Lafferty (2006) develops a method that allows for topic drift, so the probability of a token in a topic can change (slowly) through an autoregressive process. But blog data requires the possibility of rapid change; “Benghazi” did not occur in the corpus before September 11, but thereafter was a highprobability token. We develop a dynamic version of a topic model described in Yin and Wang (2014). The way we infer topics allows for both slow drift and the sudden appearance of new words, and even new topics, over the course of the year.
There is a second source of information in the blog data that previous dynamic topic models cannot utilize. It is the links between blogs, which prompt a network model. Here a blog is a node, and a hyperlink between blogs is an edge. We use an exponential random graph model (Holland and Leinhardt, 1981; Wasserman and Pattison, 1996)
to estimate the probability of an edge through a logistic regression on predictors that include node characteristics and other explanatory variables. This framework can be combined with clustering methods to perform
community detection, where a community is a set of nodes that are more likely to create edges among themselves than with other nodes.There are a number of recent methods for community detection. One is a family of algorithms that use modularity optimization (Newman and Girvan, 2004)
. But the corresponding models are not parametric and do not support Bayesian inference. A popular alternative is the latent space model of
Hoff et al. (2002). It estimates nodel locations in an unobserved space which then defines the community structure; but it is too computationally demanding for the large blog posts data set.We prefer the stochastic block model of Snijders and Nowicki (1997)
. Stochastic block models place nodes into latent communities based on the observed pattern of links between nodes, which are modeled using independent Bernoulli random variables. It has been extended as the mixed membership block model
(Airoldi et al., 2008), which allows nodes to be members of more than one community. In that spirit, the model developed in this paper keeps the stochastic block modeling framework, but permits nodes to have idiosyncrasies in their connection patterns that are not solely due to community membership, but also reflect node covariates (in this application, the degree of the blogs’ interests in specific topics). Shared community membership increases edge formation probability, and nodes in different communities that have shared topic interests also have elevated probabilities of linking. A stochastic block model can be easily expressed within an exponential random graph modeling frameworkCombining topic information and linkage information through the topic interest blocks is our key methodological contribution in this article. A secondary contribution is extending the topic model of Yin and Wang (2014) into a dynamic topic model. Researchers have started to develop models that combine network analysis and topic analysis, mostly in the context of static networks. Chang and Blei (2009) describes a relational topic model in which the probability of links between documents depends upon their topics and applies it to two datasets of abstracts and a set of webpages from computer science departments. Ho et al. (2012) applies such methods to linked hypertext and citation networks. Wang et al. (2011) develops a model for the case in which there are noisy links between nodes, in the sense that there are links between documents whose topics are not related. Yin and Wang (2014) does related work on clustering documents through use of a topic model. However none of these methods allow for the simultaneous modeling of dynamic topics with a community structure on the nodes.
Our model uses text and covariate information on each node to group blogs into blocks more likely to post on the same topics and link to one another. This approach expands upon community detection, but also fundamentally alters how communities are defined. We assume that if two blogs are interested in the same topics, then they are more likely to link to each other and form a community. Estimating the extent to which blogs post about the same topics helps explain community structure, above and beyond the community structure described by linkage pattern. Furthermore, integrating community detection into topic models allows the linkages to inform the allocation of topics, connecting network structure to topic structure. So inference on topic distributions is supplemented by nontext information. This results in communities that are defined both on the pattern of links (traditional community detection), as well as textual data. One consequence of this approach is that communities are more grounded in the substantive reason for any community structure, shared interest in various topics.
In particular, for the 2012 blog application, we wanted a bespoke Bayesian model that (1) allows topic distributions to change over time, both slowly and quickly, (2) classifies blogs into blocks that share topic interests and have elevated internal linkage probabilities, and (3) theoretically enable use of covariate information on blogs. This includes prestige, sociability, whether recently linked, and more. Here we extend covariates from static stochastic block models, as in
Faust and Wasserman (1992), to dynamic networks. Some covariates are fixed (e.g., topic interests) whereas others are timevarying (e.g., whether recently linked).Section 2 describes our dataset and its preparation. Section 3 gives a generative dynamic model for both the text and the network. Section 4 specifies the Bayesian prior and posterior inference algorithm used to estimate model parameters. Finally, in Section 5, we present several findings from the political blog data, and Section 6 finishes with a discussion of possible generalizations.
2 Political Blogs of 2012
Our data consists of the blog posts from the top 467 US political blogs for the year 2012, as ranked by Technorati (2002). This dataset has a dynamic network structure since blog posts often link to each other, responding to each other’s content. Additionally, the topic structure of the blog posts reflect different interests, such as the presidential campaign or sensational crime. The token usage in each topic changes over time, sometimes quite suddenly, as with the appearance of the tokens “Trayvon” and “Zimmerman”^{2}^{2}2George Zimmerman shot and killed Trayvon Martin in March of 2012. in March, 2012, and sometimes more gradually, as with the slow fade of the token “Gingrich”^{3}^{3}3Newt Gingrich gradually faded to political irrelevance after a failed presidential primary run. during the spring. Over the 366 days in 2012, a leap year, the political blogs accumulated 109,055 posts.
2.1 Data Preparation
Our data were obtained through a collaboration with MaxPoint Interactive, now Valassis Digital, a company headquartered in the Research Triangle that specializes in computational advertising. Using the list of 467 U.S. political blog sites curated by Technorati, computer scientists at MaxPoint scraped all the text and links at those sites (after declaring robot status and following all robot protocols).
The scraped text was stemmed, using a modified version of Snowball (McNamee and Mayfield, 2003) developed inhouse at MaxPoint Interactive. The initial application removed all threeletter words, which was undesirable, since such acronyms as DOT, EPA and NSA are important. That problem was fixed and the data were restemmed.
The second step was filtering. This filtering was based on the variance of the unweighted termfrequency, inverse document frequency (TFIDF) scores
(Ramos, 2003). The TFIDF score for token in blog post is(1) 
where is the number of times that token occurs in blog post , and is the number of posts in the corpus that use token . Words that have low variance TFIDF scores are such words as “therefore” and “because,” which are common in all posts. Highvariance scores are informative words that are used often in a small number of posts, but rarely in other posts, such as “homosexual” or “Zimmerman”. Interestingly, “Obama” is a lowvariance TFIDF token, since it arises in nearly all political blog posts.
Next, we removed tokens that were mentioned in less than 0.02% of the posts. This reduced the number of unique tokens that appeared in the corpus, as these were unlikely to be helpful in determining the topic token distribution across all posts. Many of these were misspellings; e.g., “Merkle” for “Merkel”, the Chancellor of Germany. Overall, these misspellings were either rare (as in the case of “MerkleMerkel”), or incomprehensible.
After all tokens were filtered, we computed the grams, starting with bigrams. A bigram is a pair of words that appear together more often than chance, and thus correspond to a meaningful phrase. For example, the words “white” and “house” appear in the blog corpus often, in many different contexts (e.g., race relations and the House of Representatives). But the phrase “White House” refers to the official residence of the president, and appears more often than one would predict under an independence model for which the expected number of phrase occurrences is , where is the total amount of text in the corpus and and are the proportions of the text that are stemmed to “white” and “house”. Bigrams were rejected if their significance probability was greater than 0.05. In examining the bigram set generated from this procedure, it appeared to be too liberal; English usage includes many phrases, and about 70% of tested bigrams were retained. Therefore we excluded all bigrams occurring less than 500 times corpuswide. This significantly reduced the set of bigrams.
After the bigrams were computed and the text reformatted to combine them, the bigramming procedure was repeated. This produced a set of candidate trigrams (consisting of a previously identified bigram and a unigram), as well as a set of candidate quadrigrams (made up of two previously accepted bigrams). These candidates were retained only if they had a frequency greater than 100. This cutoff removed the majority of the candidate trigrams and quadrigrams. The final vocabulary consisted of 7987 tokens.
It is possible to go further, finding longer grams, but we did not. However, we identified and removed some long gram pathologies, such as the one created by a blogger who finished every post by quoting the Second Amendment. There is a large literature on various gramming strategies (Brown et al., 1992). Our work did not employ sophisticated methods, such as those that use information about parts of speech. After this preprocessing complete, we had the following kinds of information:

Stemmed, tokenized, reduced text for each post, the date on which the post was published, the blog the post it was published on, and links to other blogs in the network.

Blog information, including the web domain, an estimate of its prestige from Technorati, and sometimes information on political affiliation.
From this information, we want to estimate the following:

Time evolving distributions over the tokens, where the time evolution on a token may be abrupt or gradual.

The topic of each post—our model assumes that a post is about a single topic, which is usually but not always the case (based upon preliminary work with a more complicated model).

The topic interest blocks, which are sets of blogs that tend to link among themselves and which tend to discuss the same topic(s).

The specific topics of interest to each of the topic interest blocks.

The linking probabilities for each pair of blogs, as a function of topic interest block membership and other covariates.

Posting rates, as a function of blog covariates and external news events that drive discussion.
We now describe the generative model that connects dynamic topic models with network models in a way that accounts for the unique features of this data set.
3 Model
The generative model can be described in two main phases: initialization of static quantities, such as blogs’ topic interest block membership, and generation of dynamic quantities, specifically posts and links. First, the model creates topic distributions that are allowed to change over time. Next, it generates timestamped news events for each topic. Each blog is randomly assigned to a topic interest block. With these elements in place, post and link generation proceeds. For each blog, on each day the number of posts from each topic is generated, in accordance to the topic interest block of that blog. The content of the post is generated from the dayspecific topic distribution, and links are generated so as to take account the blog’s topic interest block. We now describe each step in the generative model in more detail.
3.1 Topic and Token Generation
We begin with the topic distributions, which must allow dynamic change. For the th topic, on a specified day , we assume the token probabilities are drawn from a Dirichlet distribution prior. This set of topicspecific token probabilities is the topic distribution on day . To encourage continuity across days, we calculate the average of topic ’s topic distribution from the previous days and use it as the concentration parameter for the Dirichlet distribution from which the present day’s topic is drawn. The sampling proceeds in sequence, first calculating each topic’s concentration parameter as in 2 and then sampling each topic as in 3 and then moving to the next day. This procedure repeats for times . The topics are then distributed:
(2) 
(3) 
3.1.1 Topic Event Generation
To capture the eventdriven aspect of blog posting, we generate events which then boost the post rate on the corresponding topic. For each topic , at each time , there is some probability of an event occurring. One can choose for all , which suggests each topic has on average event every days. Alternatively, different topics can be given different daily event probabilities or one can put a prior on . Given , the daily, topicspecific event indicators are sampled as:
(4) 
When an event happens on topic , blogs with interest in topic have their posting rates increase by a factor determined by . Speculating that some topics have events which are much more influential than others, we let this multiplier be topic specific:
(5) 
3.1.2 Block and blog specific topic interest specification
[id=TRH]With our topic distributions and topic specific events generated, we can now assign blogs to topic interest blocks. We begin by defining the blockspecific topicinterests matrix , where each column indicates which of the topics are of interest to block . The first columns correspond to the singleton blocks, which are interested only in topic , topic , up through topic , respectively. The next columns define doublet blocks, which have interest in all of the possible topic pairs. The next columns correspond to blocks which have interest in exactly 3 topics, and the final column is for the block which has interest in all topics:
(6) 
To assign blogs to blocks, we sample their membership with a single draw from a multinomial distribution. This means each blog is a member of only a single block, characterized by the topic interests in the above matrix. Each block assignment is then drawn from a multinomial distribution:
(7) 
One can choose the probabilities of belonging to each block uniformly, by setting each element of where
gives the total number of blocks. Another approach is to partition the probabilities vector into the singlets, doublets, triplets, and alltopics blocks, and allocate probability uniformly to each of these categories, and then uniformly divide up the probability among blocks within each category:
(8) 
For notational convenience throughout the rest of the paper, we define to be the set of topics which are of interest to blog :
(9) 
With each blog’s topic interest indicators known, we can generate blogspecific topicinterest proportions. For example, two blogs may be in the block with interest in topic 1 and topic 2, but one may have interest proportions while the other has
. As is conventional in topic modeling, topic (interest) proportions are drawn from a Dirichlet distribution, though we make the distinction that each blog has a specific set of hyperparameters
. An individual topic interest vector is then a draw from a Dirichlet distribution:(10) 
The hyperparameters are chosen such that a blog with interest in topics 1 and 2 is likely to have most of its interest in those topics, though it allows for interest in other topics to occur with small probabilities:
(11) 
3.1.3 Post Generation
Given the blogs’ topic interest and block membership, along with the event distribution, we can now generate the number of posts a blog produces on a particular topic. Each blog may post on multiple topics, but each post is associated with a single topic. Every blog has a baseline posting rate which characterizes how active it generally is on days without events. For blog the baseline post rate is sampled from the following distribution:
(12) 
With the blog specific baseline post rate , the blog specific topic interest proportions , the topic specific daily event indicators , and topic specific post rate multipliers accounted for, we construct the expected post rate for each topic, on each blog, each day:
(13) 
Given this post rate, the count of posts about topic , on blog , on day are generated:
(14) 
In the observed data, we don’t know the post counts on each topic, but instead we know the marginal counts . These are referenced throughout the inference procedure described in section 4 and are calculated:
(15) 
With daily topic specific post counts and token probabilities available, the posts can be populated with tokens. We first sample a total number of tokens for each post. In particular, on day , the token count for post about topic on blog is sampled:
(16) 
Where is the average number of tokens over all posts. The tokens can then be sampled from the appropriate day and topic specific multinomial distribution with probability vector . This is done for all of the posts in the corpus like so:
(17) 
3.1.4 Network Generation
Finally, we generate the network of links between blogs. Rather than modeling link generation at a post level, we model it at a daily blog to blog level. Specifically, we model a directed adjacency matrix of links, with entry indicating whether any posts from blog have links to blog on day . The binary logistic regression is suitable for this scenario. We assume the link probability depends on the following factors:

is the similarity (in topic interests) for nodes and , and is in the interval [0,1], taking value 1 if and only if blogs and are in the same block.

indicates if blog has linked to blog within the last week (previous to the current time t).

is the average indegree (through time t1) of the receiving node i’.

is the average outdegree (through time t1) of the sending node i.
The first covariate is sampled and constructed in equations 611, and the other three covariates are defined and calculated as statistics of the past data . Together with an intercept, they comprise the regressors in a logistic regression for links, which can be written as in Equation 18,
(18) 
We specify a normal prior for the intercept and the regression coefficients:
(19) 
We can use the logistic function to write the probability of a link as:
(20) 
with coefficients and covariates written as:
(21) 
This form makes it more clear how the model can be cast within the ERGM framework (Holland and Leinhardt, 1981). Covariates are time dependent as in TERGM literature (Krivitsky and Handcock, 2014). An important note is that covariates depend only on past linking data, which makes this a predictive model of links. Finally, we sample each link as a single Bernoulli trial with the appropriate probability as defined in Equation 20:
(22) 
This generative model for the links can be thought of as a variant of stochastic block modeling (Snijders and Nowicki, 1997), where block membership is “fuzzy”. In our model, while members of the same block will have the highest probability of linking with other members of the same block, individuals who share similar topic interests, but do are not in the same block are more likely to link than individuals who share no topic interests. This allows for a pattern of linkages that more accurately reflect the empirical phenomena of topic based blog hyperlinks.
At the end of data generation we have and giving the topic interest set and topic interest proportions, respectively, for blog ; matrix with daily topic specific event indicators; array with entry giving the number of posts about topic on blog at time ; array of daily topic specific token probabilities; and multidimensional object containing the count for each token in the th post about topic on blog at time .
With a theoretically justified data generating mechanism in place, we proceed to Section to “invert the generative model” and derive posterior inference for the parameters of interest.
4 Estimation
As our dataset of blog posts consists of posts, time stamps, which blog posted each post, and each post’s links to other blogs, our inferential model needs to estimate a number of quantities. This section gives the details of how we estimate quantities of interest from the data and how we specified our priors. The notation is dense, so the following guide is helpful.

Each post’s topic assignment. We observe the content of each post, but do not know the topic assignment of each post. This must be inferred. We denote this estimate as for post .

Topic distributions. We do not know what the content of each topic is, or how each topic changes over time. We use for the topic tokendistributions matrix, and for specific topics, we denote this as .

Events and Post Rate Boosts. Events are not observed and must be inferred. This matrix is E. The eventcaused, topicspecific boosts in post rate are also inferred.

Blog specific parameters. A blog’s average post rate and topic interests must be inferred. The blog average post rate is denoted , and the topic interest proportions is a vector of length , denoted .

Blogs’ block membership. A Blog’s block membership is inferred using the linkage pattern and topic assignments of each of the blog’s posts. The th blog’s block membership is denoted as and its corresponding topic interests indicator vector is .

Network parameters. The network parameters govern the probability of linkage. These are depend upon block membership, lagged reciprocity, indegree and outdegree through a logistic regression whose coefficients must be estimated. These five network parameters (including intercept) are denoted and , respectively.
4.0.1 Hyper Parameter Specification
The model requires that several parameters be specified a priori. In this subsection we describe these hyper parameters in general terms, while in section 5.1, we show which specific values we used to analyze the political blog data. The first hyper parameter is , the total number of topics. In principle, one could place an informative prior on the number of topics and use the posterior mean determined by the data. This is, however, computationally cumbersome and so we make the decision to specify the number of topics in advance This approach is used in Blei and Lafferty (2006) and Blei et al. (2003) for the Latent Dirichlet Allocation models . One can use penalized likelihood as a selection criterion, as described in Yin and Wang (2014), or an entropy based criterion, such as the one described in Arun et al. (2010). We chose the number of topics by running models with different values of and selecting the number of topics using the entropy based criterion of Arun et al. (2010).
The time lag for topic dependency needs to be specified. This time lag determines the scale of the topics, and has units in number of days. This determines how long tokens remain in a topic, and can be conceptualized as a smoother over the time changing vocabulary. Smaller values of will produce more varied topic distributions over time, while larger values will reflect slower shifts in the topic content. For the node specific parameters, only , the Dirichlet concentration parameter on the topics which are of interests to a block, is needed. This parameter governs how often blogs are allowed to post outside of topics they are interested in, with lower values allowing for more out of interest posting, and higher values corresponding to restricted topic interests. Finally, for any reasonable number of topics, a restriction on the block structure is required to ensure computational feasibility. For an unrestricted block structure with topics, the total number of possible blocks that must be evaluated is , which is computationally intractable for moderate . In this paper, we restrict blocks to have 1, 2, or 3 topic interests, and allow one block to have interest in all topics. Finally, we specify the expected number of nonzero topic interest blocks using the prior .
4.1 A Simple Data Augmentation
While the generative model assumes Poisson distribution on post counts
, we rely on a data augmentation for the inference procedure. Because counts of posts on each blog each day are already known, we augment the generative model with latent variables which instead tell the latent topic assignment of post . We can then rewrite the Poisson likelihood as a multinomial likelihood with . This reformulation enables use of the topic assignment inference algorithm from GSDMM.4.2 Metropolis Within Gibbs Sampling
We use a Metropolis within Gibbs sampling algorithm Gilks et al. (1995) to obtain posterior distributions for the parameters defined in the generative model. This approach consists of four stages:

Each day , for each blog , sample a topic assignment for each post and update the matrix of daily topic specific tokendistributions .

For blogs, update topic interest proportions (), and base rate for posting (). For events, update the event matrix E
, and activation level parameters
. 
Update the network parameters, i.e., and .

Update each blog’s block assignment and corresponding topic interest indicators .
4.3 Topic Modeling and Post Assignment
Both posts’ topic assignments as well as the topic distributions themselves are unobserved and must be inferred. A preferred algorithm for inferring post topic assignment and topic tokendistributions would first assign each post a single topic, and second have some flexibility in collapsing topic distributions together.
To those ends, we adapt the Gibbs Sampler for the Dirichlet Mixture Model (GSDMM) of Yin and Wang (2014). As originally proposed, the GSDMM classifies a set of documents into specific topics. The tokens of a post are assumed to be generated from the topic specific multinomial which that post was assigned, and many tokens may be instantiated in multiple topics (e.g., common words such as “therefore”). The assignment of each post to a single topic differs from the Latent Dirichlet Allocation model, which models documents as mixtures across a number of topics. GSDMM estimates the probability that a document is about topic , given the current topic vocabulary distribution, as
(23) 
where is the number of posts currently assigned to topic (not including the topic assignment of post ), is the number of occurrences in post of token , and is the number of occurrences of token in topic (not including the content of post ). The
controls the prior probability that a post is assigned to a topic; increasing
implies that all topics grow equally likely. The relates to the prior probability that a token will have relevance to any specific topic; increasing results in fewer topics being found by the sampler. Finally, is the number of posts in total, and is the size of the vocabulary.As originally proposed by Yin and Wang (2014), GSDMM is a static model. We modify it by allowing to vary over time. For readability, we suppress the subscripts and denote the specific post by . We define
(24) 
to be the number of posts assigned to topic in the interval from to , not including post by blog at time . Also we let
(25) 
be the number of times that token occurs in topic in the interval from to , not including post . This defines a sliding window that allows the sampler to use information from the recent past to infer the topic to which a post belongs, while allowing new tokens to influence the assignment of the post at the current time point. The probability of assigning post to topic is then:
(26) 
where is the number of posts within the lag window. Note that (26) does not use information about the blog that generates the post. So the final step is to incorporate the tendency of blog to post on topic at time , using the Poisson rate parameter in (13). Using the normalized pointwise product of conditional probabilities, the final expression for the probability that post (i.e., ) belongs to topic is
(27) 
To reduce computation, we approximate with , as clarified in the data augmentation above.
The topic assignment of a post can now be Gibbs sampled using equation (27). The sampler assigns the first post to a single topic, updates the topictoken distributions with the content of that post, then continues to the next post, and repeats. At each time point, the sampler sweeps through the set of posts several times so that the topic assignments can stabilize. The preferred number of sweeps depends on the complexity of the posts on that day, but it need not be large. After some exploration, this study used 10 sweeps at each time point per iteration.
In summary, after topic assignment has been completed for a given day , all posts within that day will have a single topic assignment. Moving to day , the topic assignment for all posts that day will utilize the information from day through that day’s topicspecific, tokendistribution estimator specified in Equation (25). Once the topic assignment estimator reaches the final day , all posts will have assigned topics, and all topics will have a time varying tokendistribution. The post specific topic assignments are then used in the next step of estimating blogs’ topic interest vectors.
4.4 Node Specific Parameters and Event Parameters
Once posts are assigned to topics, the next step is to update the node specific parameters, specifically the blog topic interest vector and the blog posting rate .
The topic interest vector is updated in a MetropolisHastings step. As is standard in MH, we specify a proposal distribution, a likelihood, and a prior. The proposal is a draw from a Dirichlet distribution with , where is the total number of posts generated by node . The likelihood is
(28) 
where is the Poisson likelihood from equation (14) representing the day specific number of posts on blog assigned to each topic. Note the dependence of on comes through equation (13). A hierarchical prior is used, which is Dirichlet(), where the parameters are defined by the current block assignment of node , as in equation (11). This step requires estimates of each blog’s block assignment, which will be described later.
The th blog’s posting rate is also updated using a MetropolisHastings step, where the proposal distribution is a Normal truncated at 0, with mean equal to
and standard deviation equal to
. The likelihood evaluated is the same as in equation (28). The prior is a univariate Normal truncated at 0 and with mean and variance. Truncated normal distributions are used to uncouple the mean and the variance.
Next to update are the event matrix and activation boost parameters . The event matrix is updated with a series of Metropolis steps for each time point and topic. The proposal is simply if and if . The likelihood is the same as equation (28), except, for each topic at each time , the product of Poisson densities is over the blogs . The prior is simply a Bernoulli with parameter .
Each activation parameter is updated with a MetropolisHastings step, where the proposal is a truncated normal at 0 with mean and standard deviation . Again, the likelihood is similar to (28), in that it is the product indexed over time of the Poisson densities for every blogs’ number of posts in topic . Unlike the original likelihood, however, the second product is over blogs . The prior distribution on is a normal truncated at 0 with mean and standard deviation .
4.5 Network Parameters
The network parameter set consists of the vector as defined in equation (18). Each network parameter can be sampled using a Metropolis within Gibbs step. Specifically, the Bernoulli likelihood portion (equivalently a logistic regression) can be expressed as:
(29) 
To update each parameter, one conditions on all other pieces of information in the model. Proposals are normal with mean set to the current value of the parameter, and a standard deviation specific to the parameter, while the priors are normal with a given mean and standard deviation. Note here that this sampling relies on estimates of the block membership of each blog.
4.6 Block Assignment
Previously we described the sampling routine for posttopic assignment, topic tokendistributions, node specific parameters, event matrix and boosts, and network parameters. The blog specific block memberships remain to be estimated. Due to the complexity of their dependence on many other components of the model, we describe block estimation last.
Recall that a blog’s block describes two things. The first is the set of topics the blog will be most likely to post on. The second is a blog will tend to link more to other blogs that are within the same block as it is. Therefore, block assignment for a given node is informed by several pieces of data. These are its topic interests, the network parameters, other blog’s block memberships, and the observed links. One assumption of our model is that a node’s network position and topic interest are conditionally independent given block assignment, which in turn makes the sampling of a block assignment considerably simpler. After simplification, a node’s potential block assignment is informed by the number of nodes already assigned to each block. Ultimately, the probability that a node will be assigned to the th block is proportional to
(30) 
where is the number of nodes assigned to block , not including node , is related to the prior probability of being assigned to any block (analogously to in the topic model), is the complete set of network parameters, is the number of blocks with nonzero membership while node is being considered for potential assignment to block , is the prior number of blocks expected to exist, and is the set of block assignments with the th blog’s block assignment removed. As such, the first term acts as a penalty term on the number of blogs in any given block, the second term is the Bernoulli likelihood (logistic regression) of observed links given other block assignments and network covariates, and the third term is the probability of blog ’s topic interest vector given the considered block assignment. Finally is the Poisson probability of given and acts as a penalty term for the number of blocks with nonzero membership.
To elaborate briefly, the first term in (4.6) and the final term in (4.6) together act as tunable priors on the distribution of sizes of blocks, as well as the number of nonempty blocks. In the generative model, we specify individual priors on the probability of membership in each block. However, for any reasonable , the total number of blocks exceeds the number of blogs to assign to blocks. So the block assignment sampler accounts for blocks that have no members. This is a desirable feature since the analyst need not specify exactly how many blocks are instantiated in the model. As compared to the number of possible blocks, the number of nonempty blocks is a rare event. This enables use of the Poisson distribution as a reasonable (and computationally feasible) approximation for the number of nonempty blocks.
Due to the computational intensity of computing the block specific probabilities, we restrict the number of topics in which a block can have interest. In our work, blocks may be interested in at most three topics, except for one block that is interested in all topics (to account for such blogs as The Huffington Post or The New York Times’s blog). Furthermore, during sampling, we restrict the blocks considered for a given node by only considering blocks that have topics for which node generated at least one post. These restrictions change the normalizing constant, though the relative probabilities of the blocks considered remains the same.
4.7 Summary of Estimation
The primary goals of the estimation routine are to obtain posttopic assignments, and then topic tokendistributions, as well as blog specific block assignments. These are the main parameters of interest in our application, as they describe the dynamic nature of each content topic and the range of interests and communications each blog has respectively. Alongside this information, we also estimate several other parameters such as those which govern linkage formation, which topics are active when (via events), topic postrate boost parameters, and blogs’ topic interests. While each of these are informative in their own right, in our application below, we chose to focus on the topic distributions and the block assignments of each blog.
5 Results
5.1 Prior Choices
Changes in the topics for this dataset are expected to be slow, aside from sudden events that abruptly add new tokens (e.g., “Benghazi” or “Sandy Hook”). Therefore, we used a lag parameter of 62 days to capture gradual drift in the topics over time. Specifically, the distribution over tokens for each topic was estimated based upon a sliding window for the preceding 62 days. Within that window, all posts had equal weight.
To determine the number of topics, we used the criterion developed by Arun et al. (2010). The goal of this criterion is to determine the number of topics that lead to the smallest value of the criteria. It is important to note that this criteria is only based off of the topic distribution over the posts, and does not take into account the network structure. This is, of course, a limitation of this criteria and a suggestion for further research. Figure 1 shows the criterion curve, which considers fitting anywhere between 1 and 30 topics. The curve has its minimum at 22, and thus our study fixes the number of topics to be 22.
Once the number of topics is established, the restrictions on the blocks and the parameter , as introduced in equation (1), can be set. Recall that each block may only be interested in 1, 2, 3 or all topics. Finally, , the out of block interest parameter which governs the blogs ability to post on out of interest topics, was set to 50, to allow some freedom for blogs to post on topics outside of their block’s interests, but nonetheless mostly focus on the block’s interests.
The network model specified an edge parameter, a mean indegree and outdegree parameter, a 7day lag parameter, and a block membership parameter. The edge parameter acts as the intercept for the network model. Mean indegree and outdegree are nodal covariates consisting of the average daily indegree and outdegree for each node. This allows modeling of differentially popular blogs. Finally, to add in temporal dependency, the 7day lag is an indicator function that takes the value 1 if and only if the pair of blogs has been linked within the previous 7 days, and is otherwise 0 (this captures the fact that bloggers sometimes have debates, which produce a series of links over a relatively short period of time). Vague priors were set for each of the network model parameters; all were normals with mean 0 and standard deviation 1000. The proposal standard deviation was set to 1 for the edge parameter, and to 0.25 for each of the other parameters in the network model.
For the topic model the and parameters were both set to 0.1. The prior for the average post rates in equation (2) was a truncated normal at 0, with mean 4 and standard deviation 1000. The prior for topic activation parameters in equation (2) was set as a truncated normal at 0 with mean 0 and standard deviation 1000, and a proposal standard deviation of 0.5.
Additionally, 25 was set as the prior mean number of blocks (), and the prior tendency for block membership was set to . The prior probability of topic activation was set to 0.2.
The sampler ran for 1000 iterations. To ensure mixing for the network parameters, at each iteration the network parameters were updated 10 times. During each iteration, there were 10 subiterations for the topic model and 10 subiterations for the block assignments. The first 100 overall iterations were discarded as burnin, the remaining 900 were thinned at intervals of 10.
5.2 Findings
The sampler converged to stationarity quickly in every parameter. To assess the mixing of the post to topic assignment, and of the blogger to block assignment, we calculated Adjusted Rand Indices (Hubert and Arabie, 1985; Steinley, 2004) for each iteration compared to iteration . The post to topic assignment was very stable, with a mean Adjusted Rand Index of 0.806 and standard deviation 0.047. The block assignment was less stable, with a mean Adjusted Rand Index of 0.471 and standard deviation 0.031. We believe this variability is due to the fact that many bloggers tended to post on whatever news event captured their attention, making it difficult to assign them to a block with interest in no more than three topics. However, their interests were not so wide that they were reasonably assigned to the block that is interested in all topics.
All domain level parameters converged successfully. The domain rate parameter
was estimated for each domain, and the posterior means of the domain rates had a mean of 0.632 and a standard deviation of 1.67. The largest domain rate was 22.69. The distribution of domain post rates was highly skewed, with few blogs having a very high average post rate, and most blogs having a lower post rate.
The topic specific activation parameters converged successfully. Information on the posterior means and standard deviations is in Table 3, and were calculated after the topics had been defined. The topics Election and Republican Primary have the greatest posterior means, which suggests that these topics were more event driven than other topics.
5.2.1 Topic Results
The topic model found 22 topics, each of which had distinct subject matter. Table 1 contains the topic titles and total number of posts in each topic, as well as the three tokens that have the highest predictive probability for that topic over all days. Predictive probability was calculated using Bayes’ rule:
(31) 
Table 2 contains the five most frequent tokens in each topic over all days. Topics were named by the authors on the basis of the most predictive tokens as well as the most frequent tokens over all days. Some of these tokens may seem obscure, but in fact they are generally quite pertinent to the identified topics.
Topic Name  # of posts  Highest Specificity Tokens  

1  2  3  
Feminism  3971  russel.saunder.juli  circumcis  femin 
Keystone Pipeline  4422  loan.guarante.program  product.tax.credit  tar.sand.pipelin 
Birth Control  2703  contracept.coverag  birth.control.coverag  religi.organ 
Election  14713  soptic  cheroke  eastwood 
Mortgages  2130  estat  probat  fiduciari 
Entertainment  10555  email.read.add  olivia  free.van 
Middle East  6068  mursi  morsi  fatah 
LGBT Rights  5425  anti.gay.right  support.equal.marriag  equal.marriag 
Sensational Crime  6423  zimmerman  lanza  mass.shoot 
Technology  3230  mail.feel.free  pipa  ret 
Supreme Court  1767  commerc.claus  bork  chief.justic.robert 
Bank Regulation  5222  volcker  dimon  libor 
National Defense  1977  iaea  iranian.nuclear.weapon  warhead 
Republican Primary  9351  poll.mitt.romney  nation.popular.vote  romney.lead 
Voting Laws  7865  ohio.secretari.state  voter.registr.form  hust 
Political Theory  1448  bylaw  rawl  sweatshop 
Eurozone  1832  standalon  troika  ecb 
Taxation  8435  tax.polici.center  health.care.spend  top.tax.rate 
Diet and Nutrition  3057  spielberg  harlan  calori 
Education  2909  chicago.teacher.union  chicago.public.school  charter.school 
Global Warming  2205  arctic.sea.ice  sea.ice  sea.level.rise 
Terrorism  3347  kimberlin  broadwel  assang 
Topic Names  Most Frequent Words  

1  2  3  4  5  
Feminism  women  peopl  dont  person  life 
Keystone Pipeline  energi  oil  price  compani  industri 
Birth Control  right  state  law  marriag  women 
Election  obama  romney  peopl  presid  polit 
Mortgages  case  court  bank  judg  attorney 
Entertainment  peopl  dont  good  work  game 
Middle East  israel  islam  american  peopl  countri 
LGBT Rights  gay  peopl  marriag  homosexu  support 
Sensational Crime  gun  polic  report  peopl  zimmerman 
Technology  compani  googl  appl  user  
Supreme Court  law  court  state  case  constitut 
Bank Regulation  bank  market  money  price  compani 
National Defense  iran  militari  israel  nuclear  obama 
Republican Primary  romney  republican  obama  poll  vote 
Voting Laws  state  vote  elect  voter  counti 
Political Theory  libertarian  peopl  right  state  govern 
Eurozone  bank  debt  economi  rate  govern 
Taxation  tax  state  govern  cut  obama 
Diet and Nutrition  peopl  dont  govern  polit  work 
Education  school  student  teacher  educ  state 
Global Warming  climat  climat.chang  temperatur  scienc  scientist 
Terrorism  report  govern  attack  inform  case 
Topic  Posterior Mean  Standard Deviation 

Feminism  0.0034  0.0029 
Keystone Pipeline  0.0037  0.0017 
Birth Control  .00001  .00003 
Election  0.3917  0.0249 
Mortgages  .00001  .00005 
Entertainment  0.0018  0.0014 
Middle East  0.0378  0.0129 
LGBT Rights  0.0028  0.0026 
Sensational Crime  0.0208  0.0085 
Technology  0.0012  0.001 
Supreme Court  0.0022  0.0032 
Bank Regulation  0.0021  0.0021 
National Defense  0.0014  0.0013 
Republican Primary  0.2267  0.0188 
Voting Laws  0.0375  0.0095 
Political Theory  0.0012  .00001 
Eurozone  0.0001  .00002 
Taxation  0.0135  0.0084 
Diet and Nutrition  0.0804  0.0139 
Education  0.0011  0.0012 
Global Warming  0.0019  0.0011 
Terrorism  0.0022  0.0019 
It is beyond our scope to detail the dynamics of all 22 topics. However, a close look on one topic, Sensational Crime, shows the kind of information this analysis obtains. The posts about Sensational Crime largely concerned four events: the shooting of Trayvon Martin in February, the Aurora movie theater shooting in July, the Sikh Temple shooting in August, and the Sandy Hook shooting in December^{4}^{4}4Trayvon Martin was a young African American man shot by George Zimmerman, in what he claimed to be an act of self defense, while Martin was walking in Zimmerman’s neighborhood. The Aurora theater massacre was a mass shooting at a movie theater in Aurora, Colorado. The Sikh Temple shooting was a mass shooting at a Sikh temple in Wisconsin. The Sandy Hook massacre was a mass shooting at an elementary school in Connecticut..
To illustrate how the salience of a token changes over time, we use a weighted frequency proportion which is equal to:
(32) 
where is the frequency of the token in the th topic’s distribution at time . This weighted frequency can be interpreted as the proportion of topic specific tokens at time that is taken up by token
, and is useful in this context as many of the tokens are shared at high frequency between topics (such as ”people”) and are therefore uninformative. So this quantity tracks the topic specific information of a token over time. Recall that the topicspecific tokendistributions are computed over the past 62 days, which accounts for the smoothness of the curves. The gray shading around each curve represents the 95% Bayesian credible interval.
Figure 2 presents the weighted frequency curves for tokens specifically related to the shooting of Trayvon Martin. The plot shows three interesting features. First, the prevalence of all of the tokens does not spike up at the time of the shooting (2/26/2012), but rather at the time of Obama’s press statement regarding the shooting. Second, the term “zimmerman” dominates the tokens, and in fact is the most prevalent token in the whole of the Sensational Crime topic from March 22 to July 16th. The gap in prevalence between the tokens of “zimmerman” and “trayvon” or “trayvon martin” is also interesting, suggesting that in this case, media attention was on the perpetrator rather than the victim.
Figure 3 tracks the major events in the Sensational Crime topic for the entire year. Notably, media focus is never as strong on the tokens related to the events as it is on “Zimmerman” specifically. Rather, the usual top terms over the course of the year are “police” and “gun”. Also notable is the lack of events in the later part of the year after the media attention on the Sikh Temple Shooting receded.
5.3 The Network Results
[id=TRH]The use of network data improves the topic discovery, based on a comparison to analysis of the same dataset by citepospisil that used only the text. However, this analysis also enables us to use the topic modeling to improve community detection, through the incorporation of the interestgroup block structure.
Table 4 shows the posterior means of the network parameters with 95% credible intervals. The edge parameter posterior mean indicates that the network is rather sparse at most time points. Interestingly, the 7day lag parameter was negative, suggesting that blogs which were recently linked were less likely to link in the near future. There are two plausible explanations for this finding. First, the linkage dynamics may not be driven by recent links, but rather the links are a consequence of the events taking place. An upsurge in linking when an event occurs is followed by a decrease in the number of links as the event fades out of the news cycle. Second, if linking is done as part of a debate, then once a point has been made, the bloggers may not feel a need for backandforth argument.
The block parameter is strongly positive (mean = 1.058, standard deviation 0.240), suggesting that blogs which share common interests are more likely to link to each other. This is particularly important, as the block statistic was not only formed from explicit block matching, but also from blogs that did not share the same interests. The block statistic is proportional to the shared topic interests. This result directly links the network model to the topic model, and allows the analyst to make claims about the block structure as inferred from the topics.
Finally, and predictably, both the indegree and outdegree of a blog increases the probability that the block will receive links. These parameters were included in the analysis to control for the influence of highly popular blogs such as The Blaze and The Huffington Post.
Parameter  Posterior Mean  95% CI 

Edges  8.524  [8.539, 8.513] 
7 day lag  0.163  [0.198, 0.131] 
Block  1.058  [0.638, 1.485] 
Outdegree of Receiver  0.330  [0.329, 0.332] 
Indegree of Receiver  0.497  [0.496, 0.499] 
We can examine the link dynamics within a topic block. There were 21 blogs whose maximum posterior probability of block assignment placed them in the block that was only interested in the Sensational Crime topic. Only 2 of these 21 blogs received any links over the course of the year, and only 1 received links within the block (
legalinsurrection.com). While this runs counter to the idea that they form one block, recall that blogs are also more likely to link to blogs that share some of the same topic interest. There are a total of 62 blogs to which members of the Sensational Crime block link, and 15 of these blogs receive approximately 90% of the links. As such, the Sensational Crime topic block appears to be a set of “commenter” blogs that react to posts that are posted on larger blogs. Our model allows the analyst to isolate the blogs that post on a particular topic, to get a better idea of the linkage dynamics around important events. As an example, we describe how the linkage pattern changes around the occurrence of Barack Obama’s speech regarding the shooting of Trayvon Martin, and also following the Aurora shooting.Figures 4 and 5 show the link structure from the blogs in the Sensational Crime block to other blogs. The data are aggregated over fifteen days. Figure 4 pertains to the days before President Obama’s press conference regarding Trayvon Martin on 3/23/2013, and Fig. 5 pertains the days following his remarks. Figure 6 pertains to the period immediately before the Aurora shooting on 7/20/2012, and Fig. 7 pertains to the period immediately after.
To improve interpretability, only a subset of blogs and links are plotted. Specifically, blogs that were assigned to the block interested only in Sensational Crime, and who posted during the specified time frame, are plotted. Additionally, blogs who are part of the 15 blog subset that received 90% of the links from the Sensational Crime block, and which received links within the timeframe, are plotted. Also, links generated from blogs in the Sensational Crime block to other members of the same block, or to other blogs, are plotted. Links emanating from the 15node subset are not plotted. These plotting constraints help enable us to discern and interpret the community structure that formed in the discussion of these events.
Comparing Figs. 4 and 5 shows that the community structure seen in the linkage pattern did not change much as a result of the press conference in which President Obama remarked that if he had a son, he would resemble Trayvon Martin. Remarkably, there was also no net increase in posting rates. It is known that there was a flurry of posts at this time, and it turns out that uptick was allocated to the Election block, as people speculated on how his remarks would affect the 2012 presidential election.
The patterns surrounding the Aurora shooting (Figs.6 and 7) are more clear. The community structure in the discussion is essentially the same, but the amount of traffic increases conspicuously. Specifically, the number of links in the 15 days before the shooting was 197, but afterwards it was 427. Linkage rates especially increase from gunwatch.blogspot.com. In general, this agrees with the conclusion that the methodology is able to detect stable communities whose linkage rates are driven by news events.
To further illustrate the findings of the network model for a different block, we now present examples from the Election block. There are 52 blogs that the model assigned to the block whose only interest was the presidential election. Of these 52 blogs, 33 blogs linked to or received links from other blogs within this same block. And of these 33 blogs, 12 were the recipients of all links. We use random walk community detection (Pons and Latapy, 2006) upon the Election block to show that the model can extract meaningful subnetworks for use in secondary analyses.
Figure 8 shows the community substructure for the Election block aggregated over the entire year. Random walk community detection indicates that seven communities optimized modularity, but two communities contained the majority of the blogs. As such, for interpretability, only these two communities are shown. The modularity of this partition is 0.49, and a 10,000 sample permutation test of the community labels indicated that this value of modularity was in the 99th percentile (the greatest modularity found in the permutation test was 0.317). This result indicates that the model found meaningful community structure, rather than sampling variability.
Table 5 contains the blog labels for the Election block network. Examination of the block structure shows that the majority of the blogs partitioned into one of two communities. Based on Technorati ratings, the community plotted as circles in Figure 8 is politically conservative, while the other community plotted as squares is liberal. This separation of the two ends of the political spectrum has been found before in blogs (Lawrence et al., 2010). There is little communication between the two communities, but a lot of communication within those communities. Interestingly, both communities sent many links to blog 31, which was allocated into a distinct community that it shared with blogs 15, 4 and 28. Blog 31 is mediaite.com, a nonpartisan general news and media blog, and the pattern of links from both partisan communities suggests that mediaite.com acts as a common source of information.
Label  Blog  Community 

1  afeatheradrift.wordpress.com  1 
2  atlasshrugs2000.typepad.com  3 
3  bleedingheartlibertarians.com  1 
4  brainsandeggs.blogspot.com  2 
5  citizentom.com  3 
6  crethiplethi.com  3 
7  crookedtimber.org  1 
8  davedubya.com  4 
9  dogwalkmusings.blogspot.com  5 
10  driftglass.blogspot.com  1 
11  greatsatansgirlfriend.blogspot.com  3 
12  hennessysview.com  6 
13  joshuapundit.blogspot.com  3 
14  marezilla.com  3 
15  mediabistro.com  2 
16  michellesmirror.com  3 
17  nomoremister.blogspot.com  1 
18  ochairball.blogspot.com  7 
19  patriotboy.blogspot.com  1 
20  righttruth.typepad.com  3 
21  rightwingnews.com  3 
22  rogerailes.blogspot.com  1 
23  rwcg.wordpress.com  1 
24  sultanknish.blogspot.com  3 
25  tbogg.firedoglake.com  1 
26  thecitysquare.blogspot.com  3 
27  therightplanet.com  3 
28  thoughtsandrantings.com  2 
29  varight.com  3 
30  blogs.suntimes.com  1 
31  mediaite.com  2 
32  rightwingwatch.org  1 
33  patdollard.com  3 
6 Conclusion
In this manuscript we present a [id=TRH]novelBayesian model for analyzing [id=TRH]a large dataset of political blog posts. dynamic text networks. This model links the network dynamics to topic dynamics through a block structure that informs both the topic assignment of a [id=TRH]postdocument and the linkage pattern of the network.
A major feature of our model is that the block structure enables interpretable associations among topics. For example, there is a twotopic block whose members are interested in both Election topic and the Republican Primary topic, but there is no block whose members are interested in just the Supreme Court and Global Warming. That pattern of shared interest conforms to what one would expect.
Another feature of our model is the flexibility of the network model. This analysis uses a limited set of predictors, but the ERGM modeling framework can easily incorporate additional covariates (Robins et al., 2001) and structural features. Additionally, if one uses a maximum pseudolikelihood approach (Frank and Strauss, 1986) as a way of approximating the likelihood, then higher order subgraph terms, such as number of triangles or geometrically weighted edgewise shared partners (Hunter et al., 2008) can account for transitivity effects. Finally, while the block structure modeled in this paper was based upon similarity in topic interest, more nuanced models are possible, and these could use information on, say, political ideology, which the analysis of the Election block found to be important in predicting linkage patterns.
Finally, one major advantage of our approach to modeling this data is the nonparametric nature of the topic dynamics. By avoiding an autoregressive specification of topic dynamics, as in Blei and Lafferty (2006), topics are able to change more freely; in particular, it is possible for new tokens with high probability to emerge overnight. This is ideal for the blog data, since the blogs are often responding to news events.
Our analysis of the political blog dataset had an interpretable topic and block set, and analysis of the Sensational Crime block and the Election block reached reasonable conclusions. Specifically, the dominance of the token “zimmerman” across the year agrees with our sense of the tone and primacy of that discussion, and the spike following the Aurora shooting is commensurate with its news coverage. The Election block neatly split into subcommunities along partisan lines, which accords with previous research (Lawrence et al., 2010).
[id=TRH]Finally, the focus in this manuscript was the methodological development, but there are many political science questions that await deeper exploration. Among these are study of the matrix that models topic activation, which can be linked to specific headline events, the use of political affiliation as a covariate in the linkage dynamics, and examination of the association structure among blocks that are interested in more than one topic.
[id=TRH]While our focus was on the specific application of the political blog data, the model developed here has features that can generalize to other dynamic text networks. such as the Wikipedia and scientific citation networks. [id=TRH]Specifically, the connection of topic and link structure through a block structure allows for document content to inform the community structure of the overall network. However, each application requires some hand fitting that captures specific aspects of the data. For example, the block structure might need to be dynamic; this would make sense is scientific citation networks, since disciplines sometimes bifurcate (e.g., the computer science of 1970 has now split into machine learning, quantum computing, algorithms, and many other subfields). Also, scientific citation is strictly directional in time—one cannot cite future articles. But the Wikipedia is not directional in time; an article posted a year ago can send a link to one posted yesterday. So specific applications will require tinkering with the model described here.
[id=TRH]The work presented here suggests several avenues of future research. On the methodological side, the model can be generalized and extended in several ways. Specifically, the block membership could be considered a dynamic property, allowing blogs to change interest in topics over time. Additionally, building a dynamic model for link patterns would allow researchers to examine specific properties of links over time, rather than assuming the same link generating distribution at all time points. Finally, this model can be adapted to other dynamic text networks, and its performance should be compared to more traditional topic analysis and community detection procedures.
References
 Airoldi et al. (2008) AIROLDI, E. M., BLEI, D. M., FIENBERG, S. E., and XING, E. P. (2008). Mixed Membership Stochastic Blockmodels. Journal of Machine Learning Research, 9(2008):1981–2014.
 Arun et al. (2010) ARUN, R., SURESH, V., MADHAVAN, C. V., and MURTHY, M. N. (2010). On finding the natural number of topics with latent dirichlet allocation: Some observations. In Advances in Knowledge Discovery and Data Mining, pages 391–402. Springer.
 Blei et al. (2003) BLEI, D., NG, A., and AND M. JORDAN (2003). Latent Dirichlet Allocation. Journal of Machine Learning Research, 3:993–1022.
 Blei and Lafferty (2006) BLEI, D. M. and LAFFERTY, J. D. (2006). Dynamic topic models. In Proceedings of the 23rd international conference on Machine learning, pages 113–120. ACM.
 Brown et al. (1992) BROWN, P. F., DESOUZA, P. V., MERCER, R. L., PIETRA, V. J. D., and LAI, J. C. (1992). Classbased ngram models of natural language. Computational linguistics, 18(4):467–479.

Chang and Blei (2009)
CHANG, J. and BLEI, D. M. (2009).
Relational topic models for document networks.
In
International conference on artificial intelligence and statistics
, pages 81–88.  Faust and Wasserman (1992) FAUST, K. and WASSERMAN, S. (1992). Blockmodels: Interpretation and evaluation. Social networks, 14(1):5–61.
 Frank and Strauss (1986) FRANK, O. and STRAUSS, D. (1986). Markov graphs. Journal of the American Statistical Association, 81(395):832–842.
 Gilks et al. (1995) GILKS, W. R., BEST, N., and TAN, K. (1995). Adaptive rejection metropolis sampling within gibbs sampling. Applied Statistics, pages 455–472.
 Ho et al. (2012) HO, Q., EISENSTEIN, J., and XING, E. P. (2012). Document hierarchies from text and links. In Proceedings of the 21st international conference on World Wide Web, pages 739–748. ACM.
 Hoff et al. (2002) HOFF, P. D., RAFTERY, A. E., and HANDCOCK, M. S. (2002). Latent Space Approaches to Social Network Analysis. Journal of the American Statistical Association, 97(460):1090–1098.
 Hoffman et al. (2010) HOFFMAN, M., BACH, F. R., and BLEI, D. M. (2010). Online learning for latent dirichlet allocation. In Advances in Neural Information Processing Systems, pages 856–864.

Holland and Leinhardt (1981)
HOLLAND, P. W. and LEINHARDT, S. (1981).
An exponential family of probability distributions for directed graphs.
Journal of the american Statistical association, 76(373):33–50.  Hubert and Arabie (1985) HUBERT, L. and ARABIE, P. (1985). Comparing partitions. Journal of Classification, 2(1):193–218.
 Hunter et al. (2008) HUNTER, D. R., GOODREAU, S. M., and HANDCOCK, M. S. (2008). Goodness of Fit of Social Network Models. Journal of the American Statistical Association, 103(481):248–258.
 Krivitsky and Handcock (2014) KRIVITSKY, P. N. and HANDCOCK, M. S. (2014). A separable model for dynamic networks. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 76(1):29–46.
 Latouche et al. (2011) LATOUCHE, P., BIRMELÉ, E., and AMBROISE, C. (2011). Overlapping stochastic block models with application to the French political blogosphere. Annals of Applied Statistics, 5(1):309–336.
 Lawrence et al. (2010) LAWRENCE, E., SIDES, J., and FARRELL, H. (2010). SelfSegregation or Deliberation? Blog Readership, Participation, and Polarization in American Politics. Perspectives on Politics, 8(01):141.
 McNamee and Mayfield (2003) MCNAMEE, P. and MAYFIELD, J. (2003). Jhu/apl experiments in tokenization and nonword translation. In Comparative Evaluation of Multilingual Information Access Systems, pages 85–97. Springer.
 Moody (2004) MOODY, J. (2004). The structure of a social science collaboration network: Disciplinary cohesion from 1963 to 1999. American Sociological Review, 69(2):213–238.
 Newman and Girvan (2004) NEWMAN, M. E. J. and GIRVAN, M. (2004). Finding and evaluating community structure in networks. Physical Review E  Statistical, Nonlinear, and Soft Matter Physics, 69(2 2):026113.
 Pons and Latapy (2006) PONS, P. and LATAPY, M. (2006). Computing communities in large networks using random walks. J. Graph Algorithms Appl., 10(2):191.
 Ramos (2003) RAMOS, J. (2003). Using tfidf to determine word relevance in document queries. In Proceedings of the first instructional conference on machine learning.
 Robins et al. (2001) ROBINS, G., ELLIOTT, P., and PATTISON, P. (2001). Network models for social selection processes. Social Networks, 23(1):1–30.
 Snijders and Nowicki (1997) SNIJDERS, T. A. and NOWICKI, K. (1997). Estimation and prediction for stochastic blockmodels for graphs with latent block structure. Journal of Classification, 14(1):75–100.
 Steinley (2004) STEINLEY, D. (2004). Properties of the hubertarable adjusted rand index. Psychological methods, 9(3):386.
 Technorati (2002) TECHNORATI (2002). https://web.archive.org/web/20140420052710/http://technorati.com/.
 Wang et al. (2011) WANG, E., SILVA, J., WILLETT, R., and CARIN, L. (2011). Dynamic relational topic model for social network analysis with noisy links. In Statistical Signal Processing Workshop (SSP), 2011 IEEE, pages 497–500. IEEE.
 Wasserman and Pattison (1996) WASSERMAN, S. and PATTISON, P. (1996). Logit models and logistic regressions for social networks: I. An introduction to Markov graphs and . Psychometrika, 61(3):401–425.
 Yin and Wang (2014) YIN, J. and WANG, J. (2014). A dirichlet multinomial mixture modelbased approach for short text clustering. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 233–242. ACM.
Comments
There are no comments yet.