Word forms - not just their lengths- are optimized for efficient communication

03/06/2017 ∙ by Stephan C. Meylan, et al. ∙ 0

The inverse relationship between the length of a word and the frequency of its use, first identified by G.K. Zipf in 1935, is a classic empirical law that holds across a wide range of human languages. We demonstrate that length is one aspect of a much more general property of words: how distinctive they are with respect to other words in a language. Distinctiveness plays a critical role in recognizing words in fluent speech, in that it reflects the strength of potential competitors when selecting the best candidate for an ambiguous signal. Phonological information content, a measure of a word's string probability under a statistical model of a language's sound or character sequences, concisely captures distinctiveness. Examining large-scale corpora from 13 languages, we find that distinctiveness significantly outperforms word length as a predictor of frequency. This finding provides evidence that listeners' processing constraints shape fine-grained aspects of word forms across languages.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 7

page 8

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

Model

We define a probabilistic language model to characterize the distinctiveness of word forms in terms of their constituent sound-to-sound transitions. This model can be fit using a large written sample of a particular language. The starting point for the model is formalizing the task of the listener as a rational statistical inference.

Bayesian inference and distinctiveness

Upon hearing a string of sounds , a listener has to infer what word was intended by the speaker. This can be formulated as a problem of Bayesian inference. The listener should calculate a posterior distribution over words based on the sounds. Applying Bayes’ rule, this is given by

(1)

where is the probability of hearing if is the intended word, is the prior probability of the word intended by the speaker, and is the probability of hearing .

Assuming that sounds are produced faithfully, such that is close to for a particular string for each and close to otherwise, we obtain the approximation

(2)

which expresses the probability that the word is correctly identified as a function of its normalized frequency, , and the probability of the string in the language, . We define the distinctiveness of a word to be inversely related to : intuitively, a word that shares the same sound sequences with many other words is necessarily less distinctive.

Following this logic, we use the phonological information content (PIC) of a word to measure the distinctiveness of its phoneme-to-phoneme transitions [17], or its approximation in character-to-character transitions. Analogous to the metric of lexical surprisal (negative log conditional probability) used to measure the predictability of a word given preceding words [18, 19, 20]

, PIC is the negative log probability of the sequence of phonemes (distinct meaningful sounds) or letters in the string comprising a word. To obtain a compact representation of the probabilities of various phone or character sequences in a language we estimate an

n-phone or n-character model (analogous to an n-gram model over words [21] but computed over individual phonemes or characters) from a corpus sample. To support a stronger test of the relationship between distinctiveness and frequency, we avoid the circularity that more probable words necessarily contain more probable sequences by estimating the transition probabilities using the type inventory (unique words) in the language. For the smaller datasets (those from the OPUS corpus) we also present the results for a model with token-weighted phonological transitions, though we note that interpreting the correlations obtained from the weighted model is challenging given this circularity.

Under this model, the phonological information content of a word is defined as:

(3)
(4)
(5)

where are the phonemes or letters that comprise the sequence , is the length of , and n is the order of the n-phoneme or n-character model (n=5 in the current analyses to avoid overfitting). PIC does not explicitly account for the morphemic components of a word (sub-word meaningful units, like the English prefix un-), rather the the relative prevalence of morphemes is reflected in the character-to-character or phoneme-to-phoneme transition statistics.

“M” “O” “T” “O” Bits
Uniform Character Probabilities
4.700 4.700 4.700 4.700 47.004
5-Character Phonological Information Content Model
5.358 3.223 3.451 6.110 26.327
Table 1: Estimates of the information content of the word motorcycle under uniform character probability (proportional to word length) and under a type-weighted model that takes into account sequential dependencies up to length 5 obtained from the 25,000 most frequent words in the English Google Books (2012) corpus.

Length and distinctiveness

Intuitively, short words have more similar competitors and are hence easier to confuse with other words, while long words have fewer neighbors (Fig. 1). Per Zipf’s formulation, a signal which is too short may be ambiguous, and a listener is less likely to infer a speaker’s intended meaning [8]. We argue that it is not length per se which drives this effect, but rather the contribution of length to distinctiveness.

Figure 1: Longer words have fewer words with similar word forms (“neighbors”): the number of words within an edit distance of one phoneme decreases with the length of the word form. Data from 27,751 words in the English Clearpond dataset [22].

Word length is an important determinant of string probability: under a probabilistic treatment, a string that is one sound longer is a longer sequence of events, and hence of equal or lesser probability. In fact, length is strictly proportional to the probability (or log probability) of a string under a “monkeys-on-typewriters” model of letter-to-letter transitions such as that proposed by Mandelbrot [23] and further explored by Miller [10].

Under this naive statistical model, symbols (sounds, or written characters approximating those sounds) are equiprobable and independent: , where is the number of symbols in the symbol set of the language (its alphabet or phoneme inventory, for example 26 letters in English). If words were composed of equiprobable, independently-drawn symbols, would simply be .

However, this approximation fails to capture regularities in the lexical substructure observed in natural languages in two obvious ways. First, symbols are not equiprobable: across the word types in the English lexicon,

w is substantially less common than e. Second, the sound symbols are not statistically independent: the sound t in English is followed more frequently by i or e and very rarely—if ever—by b or g. People have rich knowledge of the relative prominence of these sequences in their respective languages—just as they have rich knowledge of inter-word statistical dependencies—and can call upon this knowledge in spoken word recognition [24, 25]. Ample psycholinguistic evidence also suggests that people are capable of using sub-word information, for example using sounds from the beginning of a word to predict possible continuations [26, 27, 28, 29].

When the probability of sub-word sequences is taken into account, sequences of the same length can vary markedly in distinctiveness: while xylophone and something are both nine letters long, the latter is comprised of significantly less common subsequences. When phonological information content is computed with a more accurate model of the structure of words, it differs from length because it captures deviations from independence (see Table 1).

Benefits of distinctiveness

Conceiving of words in terms of their distinctiveness has several benefits compared to length. First, distinctiveness (as measured by PIC) is a much more fine-grained measure of word form complexity, and generates predictions contrary to word length in many cases. A shorter word may contain a relatively low probability phoneme sequence (e.g., depth, 4 phonemes/5 letters, PIC = 17.86 bits under a type-weighted model from the Google Books 2012 English corpus), while a longer word may contain a higher probability, less informative sequence (e.g., ground, 5 phonemes/6 letters, PIC = 12.85 bits under that same model).

Second, PIC is closely related to metrics of lexical neighborhood density used in psycholinguistic models of spoken word recognition. Neighborhood density reflects how many words have a similar form to a given word; while proposals vary on how to best measure neighborhood density, they share the intuition that words with more similar word forms (or “neighbors”) are harder to recognize because there are more competitors consistent with a given signal. PIC is formally very similar to frequency-weighted neighborhood density [30], however it measures the number of competitors at each successive phone—a feature consistent with empirical results suggesting incremental phoneme-by-phoneme processing in some cases [29]. PIC thus constitutes a more detailed measure of neighborhood density than the canonical measure of Coltheart’s [31], the number of words within a edit distance of one of the target word. In particular, PIC is sensitive to competition effects from hearing the partial word form: while there are approximately 40 possible candidates upon hearing /Ti/ in thesis in a large sample of English, neighborhood density as assessed by Coltheart’s is much lower (the only competitors by this criterion are theses and Theseus). We return to the question of whether treating spoken word recognition as a purely sequential prediction task is an appropriate simplifying assumption in the Discussion.

A relationship between distinctiveness and frequency

We provide two listener-centric explanations as to why there might be an inverse relationship between distinctiveness and frequency. First, distinctiveness provides a way to measure the effort that listeners expend in comprehending speakers; total comprehension effort is minimized when the most frequent words are the least distinctive. Zipf’s original explanation for the inverse relationship between the frequency of a word and its length was based on the idea that languages are shaped by the desire of speakers to expend the least effort in producing words. If longer strings are more effortful to produce, it makes sense that they should be associated with less frequent words. However, we can imagine a similar argument being applied on the part of listeners: that languages are shaped by the desire of listeners to minimize the effort they expend in comprehending speakers. Distinctiveness more accurately indexes comprehension effort, above and beyond word length.

Second, an inverse relationship between distinctiveness and frequency can also be derived from an invariance argument. Recent work in computational psycholinguistics has discovered that a variety of syntactic phenomena can be predicted from the Uniform Information Density hypothesis: that languages are structured such that each word in an utterance provides approximately the same amount of information [32, 33, 34, 19]. A similar invariance principle that might be relevant at the level of the words themselves is Uniform Recognizability: that the probability any word is successfully recognized is approximately the same. Under Equation 2, the probability of successfully recognizing from its associated string is . Making this constant across words means that we should expect and to be directly related, and hence and to be negatively correlated.

To evaluate this hypothesis we evaluate the correlation between frequency and PIC across large corpus samples of 13 languages from three large-scale datasets: web scrapes from the Google 1T corpora [35, 36], scanned books from the Google Books 2012 corpus [37], and subtitles from the 2013 OPUS corpus [38]. We limit our analysis to languages with phonemic scripts. Following Piantadosi et al. [19], we compute correlations over the 25,000 most frequent types in each language; following that same study we additionally compute the correlation between PIC and average in-context information content (trigram surprisal) over the same set of words.

Results

We investigate the correlation between distinctiveness and frequency in large corpus samples (43m to 266b words) in 13 languages, across three large-scale datasets. In each case we compute the frequency and in-context information content (mean trigram surprisal, or negative mean log conditional probability under a trigram model) for each word, and measure its PIC under an -phoneme and -character model estimated from distinct words in the language (for additional details, see Methods). For the OPUS datasets, we also compute PIC under the analogous token-weighted models. If length is a reduced-resolution approximation of distinctiveness, then we expect an even stronger relationship between distinctiveness—as measured by PIC—and frequency across languages.

Cross-Linguistic Results

Following the methodology adopted in [19], we examine the correlation between word-level predictors (frequency and in-context information content) and metric of structural form (word length or PIC) for the 25,000 most frequent types in each language. Unlike [19], we limit our analysis to in-dictionary types, thereby excluding person names, place names, and acronyms from the analysis. We obtain a systematically stronger negative correlation between frequency and distinctiveness, as measured by type-weighted model PIC, than frequency and word length (Fig. 3). Even holding word length constant, distinctiveness explains substantial additional variance in word frequency (Fig.  2). Building the model from phonemic transcriptions, this pattern holds in 11 of 11 languages in the Google 1T datasets, 6 of 7 languages from Google Books 2012, and all 13 languages from the 2013 OPUS corpus. Building the model from raw characters—as an approximation of the phonological form—this pattern holds in all cases. In many cases (10 of 11 languages in Google 1T, 3 of 7 in Google Books 2012, and 1 of 13 in OPUS) the partial correlation of PIC and frequency—with word length partialed out—is higher than the simple correlation of frequency and word length. The obtained correlations are even stronger than those recently obtained between word length and average in-context information content [19].

Figure 2: For words with the same number of meaningful sounds (in this case 7 phonemes), distinctiveness measured with Phonological Information Content accounts for additional variance in frequency. The red line indicates the best linear fit for all words with seven sounds. Though data presented here reflect phonemic transcriptions, words are labeled in standard English form. Labeled points (black) are randomly sampled from the words in the sample with phonemic transcription of this length.

PIC computed under the token-weighted model demonstrates an even stronger correlation with frequency across the languages in the sample (Fig. 4). We caution against an overly strong interpretation of this result, however, in that this correlation is not significantly higher than that obtained when phonemic content is shuffled among word forms (maintaining length for each word) and PIC subsequently recomputed. In contrast, the correlation between frequency and PIC computed using a type-weighted model under the permuted dataset is no stronger than the correlation between frequency and length.

A similar pattern of results emerges regardless of whether the type-weighted model is computed over characters or phonemes; the sole exception is the Russian corpus from Google Books 2012 where word length is a stronger predictor when the model is computed over phonemes. However, this dataset is an outlier in two notable ways. Russian shows the lowest correlation between frequencies obtained from Google Books and OPUS (Pearson’s

= .48), as well as the lowest correlation between PIC estimates derived from Google Books and those derived from OPUS (Pearson’s = .63). Across languages, models built over phonemes and character transitions provide proportional estimates of PIC (Pearson’s between .789 and .919 across languages, median = .874). While more research is required to extend these findings beyond Germanic, Romance, and Slavic languages, Hebrew provides an important test of whether this relationship holds in languages with extensive nonconcatenative in addition to affixal morphology.

Figure 3: Frequency demonstrates a stronger negative correlation with distinctiveness (as measured by type-weighted phonological information content, or PIC) in all datasets analyzed; this difference is significant in all cases but one (Russian in Books 2012). Bars indicate Spearman’s for the two variables for the

most frequent words in each dataset. Gray lines indicate the 99% bootstrapped confidence interval. Black lines indicate the correlation with the other measure of word difficulty partialed out. After distinctiveness is partialed out, positive partial correlations are obtained for many datasets. Circles at the top of the figure indicate the number of tokens in each dataset.

Figure 4: Token-weighted phonological information content (PIC) is an even stronger predictor of frequency than type-weighted PIC. Bars indicate Spearman’s for the two variables for the most frequent words in each dataset. Gray lines indicate the 99% bootstrapped confidence interval in each case.

Relationship to Preceding Context

While Piantadosi et al. [19] found that taking into account contextual predictability / information content (in the form of mean trigram surprisal in a dataset) better predicts word length than using a simple frequency measure (operationalized as unigram surprisal, or negative log probability), we find a qualitatively different pattern of results for PIC. In the analysis above, we find that the correlation between frequency and PIC is higher than the correlation between mean trigram surprisal and PIC in all cases (Fig. 5, A). The correlation between frequency (negative log unigram probability) and PIC is also greater than the correlation between mean trigram surprisal and word length in all but one case (Fig. 5, B). We find substantially attenuated support for the principal claim in Piantadosi et al. [19], in that we find unigram frequencies better predict word length than does mean trigram surprisal (Fig. 5, C). This discrepancy, which we explore further in the Discussion, seems to reflect refinements in the list of lexical items analyzed, improvements in the data preparation and analysis methodology in the current work, and underlying issues in the construct validity of average information content for word forms in languages with rich morphology.

Figure 5: Comparison of correlations between combinations of sentence-level predictability (unigram surprisal vs. mean trigram surprisal) and word-level measure (word length or phonological information content, PIC). Bars indicate Spearman’s for the two variables for the most frequent words in each dataset. A. PIC is more strongly correlated with unigram surprisal (negative log unigram probability) than mean trigram surprisal (negative mean log conditional probability) across a broad range of languages and datasets. B. The correlation between PIC and frequency is stronger than the correlation between mean trigram surprisal and length in all but one dataset (English Google 1T). The latter is the key relationship identified in Piantadosi et al. [19]. C. Among words found in a relevant language-specific dictionary, unigram surprisal (negative log normalized frequency) is a better predictor than mean trigram surprisal of word length in most cases. Notable exceptions are English and French in Google 1T and English and German in Google Books 2012. Correlations from the OPUS datasets are omitted because these datasets are too small to reliably estimate mean trigram surprisal for most languages.

Discussion

The relationship between word length and frequency is one of the most robust empirical findings regarding the structure of natural languages. Treating word recognition in terms of Bayesian inference, we show that the requirement for successful recognition not only motivates the relationship between frequency and length, but generates even stronger predictions regarding the relationship between frequency and the probability of the phoneme sequences (and their approximation in terms of character sequences) that make up word forms. Analyses of large-scale corpora from 13 languages across three datasets (including the linguistic content of web pages, books, and movie subtitles) substantiate these claims. In this section we consider some of the implications of these results, as well as caveats and future directions for research.

Figure 6: For words that have entered English more recently, the correlation between frequency and PIC is higher than the correlation between frequency and word length. This is consistent with the hypothesis that PIC can capture finer-grained changes in word forms—which are more likely on shorter timelines–besides changes in length. Error bars show bootstrapped 99% confidence intervals.

The approach that we have taken in this paper potentially provides a link between the psycholinguistic processes at work in producing and perceiving speech and the factors that shape human languages. Previous research has demonstrated that speakers may “reduce”—underarticulate, shorten, weaken, or omit altogether—frequent or highly predictable words in specific contexts, for example using “probly” in place of “probably” in fluent speech [39, 40, 41]. On longer timescales, variants with more probable forms may become the dominant form in the language, e.g., English lunch displaced luncheon in the 19th and 20th centuries [42] (see Mahowald et al. [43] for additional examples).

PIC provides a metric by which phonemic and phonetic lexical variants may be compared while capturing variation beyond that of word length. Whereas word length can only reflect deletion and epenthesis (the insertion of phonemic material), PIC is sensitive to changes that maintain the same length, including assimilation (a sound becoming more similar to a neighboring sound), dissimilation (a sound becoming less similar to a neighboring sound), lenition (a consonant taking on more vowel-like qualities), and metathesis (transposition of proximal sounds). The change from Middle English aks to ask, for example, is reflected in a change in PIC, though both have the same length. Paradigmatic sound changes operating across many words—or indeed an entire language (e.g. phonemic mergers)—can also be characterized in terms of changes in PIC, similar to entropy-based estimates of functional load [44, 45].

The formulation of the relationship between word form distinctiveness and frequency that we present above also makes several testable predictions regarding the relationship between frequency, length, PIC, and how long a word has been in a language. First, we expect words that have recently entered a language such as neologisms and loanwords from other languages to exhibit a relatively weak relationship between frequency and their structural form, regardless of whether that structural form is characterized in terms of either PIC or length. In the case of neologisms, semantic transparency of the inputs initially determines structural form; in the case of loanwords, the it is heavily influenced by the phonological form in the language from which the word is borrowed. Second, we expect a word’s structural form comes to better reflect frequency over the course of time, in that changes that improve communicative efficiency can proceed gradually, in part because variation is constrained by the need to maintain intelligibility within a population of speakers. Third, frequency should demonstrate a stronger correlation with PIC than with length on shorter timescales, in that PIC is sensitive to more subtle changes in the word form than the change in length. For example, deletion of a phoneme often comes only after an intermediate stage where that phoneme is weakened; PIC quantitatively reflects the intermediate weakening, whereas word length changes when the phoneme is deleted entirely.

In order to test these hypotheses, we obtain the date of first appearance for 31,027 English types from the Oxford English Dictionary. We stratify these dates by century, assigning pre-700 (corresponding to proto-Old English) words to the 7th century, and compute Spearman’s rank correlation coefficient against current word frequency for the set of words corresponding to each century. The correlations we obtain between PIC and frequency are stronger than those between length and frequency for words entering English between the 13th and 20th century (Fig.  6). This pattern of results provides qualitative support for all three hypotheses: an overall weaker relationship between frequency and structural form among recent words, an increasingly strong relationship for words that have been in the language for a longer interval, and a stronger relationship between PIC and frequency for those words that have entered the language more recently. With additional datasets we hope to characterize how changes in frequency relate to specific changes in word form.

In contrast to the original formulation by Zipf which focused on minimization of speaker effort, in-context predictability as presented in Piantadosi et al. [19] and word form distinctiveness both implicate listener-oriented pressures in the relationship between frequency and word form. The two proposals make contrasting predictions regarding what factors are most relevant to generalizing the relationship, however. In-context predictability generalizes the notion of a listener’s prior beliefs regarding the probability of a given word, replacing frequency with predictability. Distinctiveness provides a more detailed characterization of word form complexity. In principle, both proposals could be true: the strongest relationship could be between in-context predictability and word form distinctiveness, though we do not find empirical support for this in the datasets analyzed here. Instead, as noted in the results we find in the current work substantively attenuated support for the relationship between word length and average in-context predictability (Fig. 5). Here we further investigate the source and significance of this discrepancy with the findings of Piantadosi et al. [19].

Using the same list of words and in-context information content estimates from Piantadosi et al. [19]

, we classify each word into one of the three categories: those found in the relevant dictionary (by testing for membership in the corresponding GNU Aspell dictionary for each language), those found in English (by testing for membership in the English Aspell dictionary for words in non-English languages), and label the remainder as out-of-dictionary. We designate those words found both in the language-specific dictionary as well as English, e.g. Spanish

pan, as in-dictionary items. The proportion of items found in each category for each language, as well as example classifications from Spanish, are presented in the inset in Fig. 8, center.

Figure 7: Piantadosi et al. [19] found higher global correlations between word length and in-context predictability (information content) as measured by mean trigram surprisal (blue bars in panel 1) than between word length and frequency (unigram surprisal), red bars in panel 1). The correlation is substantially weaker within each of three sub-groups of lexical items: types that are found in the relevant dictionary (panel 2), those found in English (panel 4), and those found in neither dictionary (panel 3).

Under this analysis, within-group correlations for all three groups are substantially lower than the aggregate correlations reported in Piantadosi et al. [19] (Fig. 7). Instead, the high global correlation between mean in-context information content and word length emerges from the inclusion of out-of-dictionary items and (for languages other than English) items from English, which are both more predictable and shorter than in-dictionary words (Figure 8

, right). The global correlations between frequency and word length found in that study were depressed by out-of-dictionary items, which were shorter—yet oddly

less frequent—than in-dictionary terms (Figure 8, left).

Because [19] represented word forms as closest ASCII equivalents (e.g., manana for Spanish mañana), some words cannot be found in the relevant dictionary. This means that for the above analysis some lexical items that should have been classified as in-dictionary have been classified as out-of-dictionary, with unknown implications for the within-group correlations. Because there is no way to systematically restore these word forms, we compare the correlation between in-context information content (mean trigram surprisal, which they also refer to as information content) and word length, as well as unigram surprisal (negative log normalized frequency) and word length using estimates from the main analysis.

Using 25,000 in-dictionary words—and respecting the original character encoding—we find that frequency is a better predictor of word length than in-context information content in most languages among those words that appear in the corresponding language’s dictionary. Key exceptions, however, are consistent with the findings of Piantadosi et al. [19]: English and French in the Google 1T and English and German in Google Books 2012 show a stronger correlation of mean trigram surprisal and word length (Fig. 5

, C). This may reflect that strong correlations only emerge with unbiased estimates of in-context information content, which may in turn only be obtained in very large datasets: English Google 1T and Books 2012 corpora are at least five times larger than the next largest dataset. In the case of Google 1T, this correlation is higher than that of frequency and PIC, leaving open the possibility that word length better reflects in-context information content while PIC more strongly reflects frequency.

Figure 8: 2-D density plots depicting the relationship between sentence-level predictor (unigram surprisal or mean trigram surprisal) and word length for the 25,000 most frequent words in 11 languages used in Piantadosi et al. [19]

. Words are stratified into three categories: those in a spelling dictionary of the target language (blue), those in a spelling dictionary for English (red), and those not in either (typically names and abbreviations, green). High correlations between mean trigram surprisal and word length, as well as low correlations between unigram surprisal emerge from an interaction between these groups. Marginal boxplots show the median, inter-quartile range (IQR), and 1.5 * IQR for each of the groups. Densities are normalized per group.

While the heterogeneous pattern of results in which structural form is more closely linked to frequency in some languages and to in-context information content in others may reflect genuine cross-linguistic differences, in the current section we argue that computing in-context information content using -gram models may be theoretically unsound in morphologically-rich languages. Consider for example that whereas an -gram model for English would have entries for a handful of forms for the verb sell (e.g., sell, sells, sold, selling), an -gram model for Spanish, a language with much richer morophology for verbs, needs to have many more entries for the corresponding verb vender owing to the combinatorial space of possible conjugations and object clitics (160 in Google Books 2012). Aggregating across inflected forms, the forms of vender and sell are approximately equally common, yet estimates of mean in-context surprisal may are more likely to be biased in Spanish given the sparsity of certain forms (e.g., véndeselas, corresponding to English imperative you sell them that). Second, if an analysis is conducted on the top most frequent word forms, high frequency lemmas take up more of the list because they are spread across a larger number of inflected entries. Depending on what parts of speech have high morphological complexity, substantive differences may emerge in the composition of the vocabulary under analysis across languages. Future work will need to investigate how morphological complexity interacts with in-context information content across languages, and whether different tokenization (segmentation) processes should be used, for example determining whether object clitics should be treated as separate words in Spanish. For now, we note that we obtain the same pattern of results for PIC and frequency when we restrict our analysis to morphologically simple word forms for those languages with morphological annotation in the CELEX database [46] (English, German, and Dutch).

We make two key simplifications in the current work with respect to the measurement of PIC: we use a relatively simple, purely sequential representation of the phonemes that comprise a word and we compute PIC with respect to the true preceding context within each word. Using a richer hierarchical representation of word form structure may account for regularities than an -phoneme model cannot, leading to a modest improvement in predicting a held-out set of words [47]. Treating spoken word recognition as sequential recognition of phonemes may also discount regularities in morophological structure, though in many cases such regularities are captured implicitly in the transition probabilities among phonemes. Adapting more elaborate models of lexical structure—both morphological and phonological—may lead to improved estimates of word form probability in a language.

With respect to the second simplification, listeners do not have access to the true identity of preceding phonemes, and instead likely marginalize over a distribution of preceding as well as following sounds within a word in the process of word recognition [48]. This may mean that listeners have comparatively peaked estimates of the preceding sounds only in the case of longer words,, such that the relationship between frequency and PIC as computed here is weaker for short words. In future work we intend to investigate whether a measure of uncertainty that takes this into account can explain additional variance in word frequency.

Data Issues

Corpus analyses of the scale used here necessarily contain some degree of noisy data. The choice of texts, methods for identifying content (i.e., excluding page numbers, tables, publisher notes etc.), optical character recognition procedure, and tokenization all influence the obtained dataset. Additional data preparation decisions such as text encoding, treatment of punctuation, and treatment of upper and lower case forms can all have pronounced effects on lexical statistics.

Despite these challenges, these datasets constitute the best resource available for computing average in-context information content for lower frequency words, given that their huge size provides better evidence of the range of contexts in which these words appear. In the course of the analysis we found two notable issues in both the Google Books 2012 and Google 1T datasets; we include a brief note on each here to alert others using this or similar corpora.

First, we found extensive evidence of issues with word segmentation in the Hebrew corpus of Google Books 2012 (for details of the rule-based tokenization procedure see Lin et al. [49]). We find a high proportion of bound morphemes (l/s, t’, l, /s, b, w, h) listed as separate lexical items in the raw data, in total accounting for approximately 20% of the total unigram probability mass.

Second, the web-based nature of the Google 1T corpora strongly influences certain average in-context surprisal estimates. For example, “Romanian” has very low trigram surprisal in the Google 1T corpus because it is overwhelmingly encountered in language lists or lists of currencies (22% of instances appear in the context “Polish Portuguese Romanian”). Furthermore, the statistical properties of the everyday language environment of speakers may differ substantially from the corpora examined here. Nonetheless, these corpora constitute the best approximation to naturalistic language use of sufficient size to support lexicon-wide analysis, particularly the calculation of mean information content for relatively rare words. As more corpora are made available, we hope to extend the analyses presented above to a more diverse set of languages.

Conclusion

The canonical inverse relationship between word length and frequency is a special case of an even broader relationship between word distinctiveness and frequency. Distinctiveness plays a crucial role in word recognition, capturing the strength of competing targets for a speech signal. Rational analysis reveals limitations on word forms: speakers can simplify and shorten words, but they are limited by listener’s requirements for distinctive word forms for successful recognition.

Methods

Datasets for Frequency and Surprisal Estimates

The Google Web 1T datasets were downloaded from the Linguistic Data Consortium [36, 35]; the Google Books 2012 datasets were downloaded from storage.googleapis.com/books/ngrams/books/ datasetsv2.html [37], and OPUS (2013) from opensubtitles.org [38]. All -grams with punctuation-only words were discarded, and punctuation appearing with other text, with the exception of apostrophes, was removed. We make the simplifying assumption that the tokenized orthographic forms correspond to psychologically salient words in the lexicon of speakers; while this assumption does not hold for all forms (e.g. German compound nouns), it holds for the vast majority of word forms in the analysis (see also our analysis of morphologically-simple forms in the Discussion, which also addresses this question). All characters were converted to lowercase using the relevant POSIX locale; US English and European Portuguese were used for English and Portuguese, respectively. In the case of Google Books 2012, records with part-of-speech tags were discarded, along with records from earlier than 1800. UTF-8 encoding was maintained throughout for all languages and datasets. Hebrew strings were represented with right-normalized forms. Counts were stored using ZS, a specialized file format for efficient retrieval of -gram counts [50].

Estimating Sentential Information Content

Following Piantadosi et al. (2013) we analyze a word list constructed from the 25,000 most frequent words in each dataset. Token frequencies were computed from the 2013 release of the the OPUS subtitle corpus. For each language, the list of unique types were filtered by those words recognized by the UNIX utility Aspell for the relevant locale. After filtering, we computed the negative log unigram probability (proportional to log normalized frequency) for each word along with the negative mean log trigram probability across contexts, following [19], where is the context for the th occurrence of and is the frequency of in the dataset.

Estimating Phonological Information Content

For the type-weighted models, a five-character transition model was estimated for each language using the 25,000 most frequent in-dictionary words also appearing in the corresponding OPUS subtitle corpus. For the token-weighted models (OPUS only), a five-character transition model was estimated using all in-dictionary tokens in the corresponding OPUS subtitle corpus. In both the token- and type- weighted cases, we also produced a five-phone transition models for all languages with the exception of Hebrew using IPA transcriptions from an automatic speech synthesizer, eSpeak. Using IPA representations for words accounts for language-specific variations in orthographic conventions. For example, written Spanish includes accents only when the placement of prosodic stress cannot be deduced from more general rules in the language. Using an IPA transcription avoids the need for developing language-specific decisions, for example deciding whether ‘a’ vs. ‘á’ should be merged or kept as separate orthographic variants in Spanish. Loan words and acronyms can greatly affect the obtained transition probabilities, especially when the transitions observed in the type-weighted model (e.g., if the transitions in “Okeechobee,” “mañana,” and “ACLU” are as heavily weighted in a phonotactic model of English as the transitions in “they” and “will”). To minimize these effects, we use only non-capitalized types present in Aspell dictionaries to build sound and character transition models for each language (with the exception of German, in which common nouns are capitalized). To avoid overfitting among higher order sequences, phone and character transition probabilities for the type-weighted models were computed with modified Kneser-Ney smoothing [51]

with interpolation on orders 3, 4, and 5 using the SRILM toolkit

[52]. Good-Turing smoothing was used for the token-weighted models. Each word’s phonological probability was calculated as the product of the probabilities of each symbol given the preceding symbol string, including a start symbol , e.g., . The probability of the end symbol was omitted in that this appreciably inflates the PIC of short words under a type-weighted model (“to” is an unlikely/high-surprisal sequence because there are many words that begin with “to”).

Code availability

Our library for the rapid cleaning, manipulation, summarization, and querying of -gram data is available at github.com/smeylan/ngrawk. Jupyter notebooks for the analyses presented here are available at github.com/smeylan/pic-analysis.

Acknowledgements

This material is based upon work supported by the US National Science Foundation Graduate Research Fellowship under grant no. DGE-1106400 and NSF grant no. SMA-1228541. Special thanks to Steven Piantadosi for sharing materials and lexical information content estimates, helpful commentary on early drafts from Terry Regier, and members of the Computational Cognitive Science Lab at UC Berkeley for valuable discussion.

References

  • Greenberg [1963] JH Greenberg. Some universals of grammar with particular reference to the order of meaningful elements. In JH Greenberg, editor, Universals of Human Language, pages 73–113. MIT Press, Cambridge, MA, 1963.
  • Evans and Levinson [2009] N Evans and SC Levinson. The myth of language universals: language diversity and its importance for cognitive science. Behav Brain Sci, 32(5):429–448, 2009.
  • Futrell et al. [2015] R Futrell, K Mahowald, and E Gibson. Large-scale evidence of dependency length minimization in 37 languages. Proc. Natl. Acad. Sci. U.S.A., 112(33):10336–10341, 2015.
  • Hauser et al. [2002] MD Hauser, N Chomsky, and WT Fitch. The faculty of language: what is it, who has it, and how did it evolve? Science, 298(5598):1569–1579, 2002.
  • Kemp and Regier [2012] C Kemp and T Regier. Kinship categories across languages reflect general communicative principles. Science, 336(6084):1049–1054, 2012.
  • Fedzechkina et al. [2012] M Fedzechkina, TF Jaeger, and EL Newport. Language learners restructure their input to facilitate efficient communication. Proc. Natl. Acad. Sci. U.S.A., 109(44):17897–17902, 2012.
  • Zipf [1935] GK Zipf. The Psychobiology of Language. Houghton-Mifflin, 1935.
  • Zipf [1949] G Zipf. Human Behaviour and the Principle of Least-Effort. Addison-Wesley, Cambridge, MA, 1949.
  • Yule [1944] G.U. Yule. The Statistical Study of Literary Vocabulary. Cambridge University Press, 1944.
  • Miller [1957] GA Miller. Some effects of intermittent silence. American Journal of Psychology, 70:311–314, 1957.
  • Cancho and Solé [2003] Ramon Ferrer i Cancho and Ricard V. Solé. Least effort and the origins of scaling in human language. Proceedings of the National Academy of Sciences, 100(3):788–791, 2003.
  • Conrad and Mitzenmacher [2004] B. Conrad and M. Mitzenmacher. Power laws for monkeys typing randomly: the case of unequal probabilities. IEEE Transactions on Information Theory, 50(7):1403–1414, 2004.
  • Piantadosi [2014] S.T. Piantadosi. Zipf’s word frequency law in natural language: A critical review and future directions. Psychonomic Bulletin & Review, 21(5):1112–1130, 2014.
  • Norris and McQueen [2008] D Norris and JM McQueen. Shortlist B: a Bayesian model of continuous speech recognition. Psychol Rev, 115(2):357–395, Apr 2008.
  • Balling and Baayen [2012] LW Balling and RH Baayen. Probability and surprisal in auditory comprehension of morphologically complex words. Cognition, 125(1):80–106, Oct 2012.
  • Hawkins [1994] JA Hawkins. A Performance Theory of Order and Constituency. Cambridge University Press, Cambrdige, UK, 1994.
  • Cohen Priva [2008] U Cohen Priva. Using information content to predict phone deletion. In Proceedings of the 27th West Coast Conference on Formal Linguistics, pages 90––98, Somerville, MA, 2008. Cascadilla Proceedings Project.
  • Levy [2008] R Levy. Expectation-based syntactic comprehension. Cognition, 106(3):1126–1177, 2008.
  • Piantadosi et al. [2011] ST Piantadosi, H Tily, and E Gibson. Word lengths are optimized for efficient communication. Proc. Natl. Acad. Sci. U.S.A., 108(9):3526–9, 2011.
  • Smith and Levy [2013] NJ Smith and R Levy. The effect of word predictability on reading time is logarithmic. Cognition, 128(3):302–319, 2013.
  • Manning and Schütze [1999] CD Manning and H Schütze.

    Foundations of Statistical Natural Language Processing

    .
    MIT Press, Cambridge, MA, USA, 1999. ISBN 0-262-13360-1.
  • Marian et al. [2012] V. Marian, J. Bartolotti, S. Chabal, and A. Shook. CLEARPOND: cross-linguistic easy-access resource for phonological and orthographic neighborhood densities. PLoS ONE, 7(8):e43230, 2012.
  • Mandelbrot [1954] B Mandelbrot. Simple games of strategy occurring in communication through natural languages. Transaction of the IRE Professional Group on Information Theory PGIT, 3(3):124–137, 1954.
  • Vitevitch and Luce [1999] MS Vitevitch and PA Luce. Probabilistic phonotactics and neighborhood activation in spoken word recognition. J Mem Lang, 40(3):374–408, 1999.
  • Luce and Large [2001] PA Luce and NR Large. Phonotactics, density, and entropy in spoken word recognition. Lang Cognitive Proc, 16(5-6):565–581, 2001.
  • Marslen-Wilson and Welsh [1978] WD Marslen-Wilson and A Welsh. Processing interactions and lexical access during word recognition in continuous speech. Cognitive Psychology, 10(1):29 – 63, 1978.
  • Marslen-Wilson [1987] WD Marslen-Wilson. Functional parallelism in spoken word-recognition. Cognition, 25(1-2):71–102, 1987.
  • Zwitserlood [1989] P Zwitserlood. The locus of the effects of sentential-semantic context in spoken-word processing. Cognition, 32(1):25–64, 1989.
  • Eberhard et al. [1995] KM Eberhard, MJ Spivey-Knowlton, JC Sedivy, and MK Tanenhaus. Eye movements as a window into real-time spoken language comprehension in natural contexts. J Psycholinguist Res, 24(6):409–436, 1995.
  • Luce and Pisoni [1998] PA Luce and DB Pisoni. Recognizing spoken words: The neighborhood activation model. Ear Hear, 19(1):1–36, 1998.
  • Coltheart et al. [1977] M Coltheart, E Davelaar, JT Jonasson, and D Besner. Access to the internal lexicon. In S Dornic, editor, Attention and Performance VI, pages 535–555. Lawrence Erlbaum Associates, 1977.
  • Genzel and Charniak [2002] D Genzel and E Charniak. Entropy rate constancy in text. Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), pages 199–206, 2002.
  • Aylett and Turk [2004] M Aylett and A Turk. The smooth signal redundancy hypothesis: a functional explanation for relationships between redundancy, prosodic prominence, and duration in spontaneous speech. Lang Speech, 47(1):31–56, 2004.
  • Levy and Jaeger [2007] R Levy and TF Jaeger. Speakers optimize information density through syntactic reduction. In B Schölkopf, J Platt, and T Hoffman, editors, Advances in Neural Information Processing Systems 19, pages 849–856, Cambridge, MA, 2007. MIT Press.
  • Brants and Franz [2009] T Brants and A Franz. Web 1T 5-gram, 10 European Languages Version 1 LDC2009T25, 2009.
  • Brants and Franz [2006] T Brants and A Franz. Web 1T 5-gram Version 1 LDC2006T13, 2006.
  • Michel et al. [2011] J Michel et al. Quantitative analysis of culture using millions of digitized books. Science, 331(6014):176–182, 2011. doi: 10.1126/science.1199644.
  • Tiedemann [2012] J Tiedemann. Parallel data, tools and interfaces in OPUS. In Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC’12), 2012.
  • Aylett and Turk [2006] M Aylett and A Turk. Language redundancy predicts syllabic duration and the spectral characteristics of vocalic syllable nuclei. The Journal of the Acoustical Society of America, 119(5):3048–3058, 2006.
  • Bell et al. [2009] A Bell, JM Brenier, M Gregory, C Girand, and D Jurafsky. Predictability effects on durations of content and function words in conversational english. Journal of Memory and Language, 60(1):92–111, 2009.
  • Gahl et al. [2012] S Gahl, Y Yao, and K Johnson. Why reduce? phonological neighborhood density and phonetic reduction in spontaneous speech. Journal of Memory and Language, 66(4):789–806, 2012.
  • oed [2016] Oed online, 2016.
  • Mahowald et al. [2013] K. Mahowald, E. Fedorenko, S. T. Piantadosi, and E. Gibson. Info/information theory: speakers choose shorter words in predictive contexts. Cognition, 126(2):313–318, Feb 2013.
  • Surendran and Niyogi [2006] D Surendran and P Niyogi. Quantifying the functional load of phonemic oppositions, distinctive features, and suprasegmentals. Amsterdam Studies in the Theory and History of Linguistic Science Series 4, 279:43, 2006.
  • Wedel et al. [2013] A Wedel, A Kaplan, and S Jackson. High functional load inhibits phonological contrast loss: A corpus study. Cognition, 128(2):179–186, 2013.
  • Baayen et al. [1995] HR Baayen, R Piepenbrock, and L Gulikers. The CELEX lexical database. Release 2 (CD-ROM), 1995.
  • Futrell et al. [In press] R. Futrell, A. Albright, P Graph, and T.J. O’Donnell. A generative model of phonotactics. Transactions of the Association for Computational Linguistics, In press.
  • Strand and Liben-Nowell [2016] J Strand and D Liben-Nowell.

    Making long-distance relationships work: Quantifying lexical competition with hidden markov models.

    Journal of Memory and Language, 90:88 – 102, 2016.
  • Lin et al. [2012] Yuri Lin, Jean-Baptiste Michel, Erez Lieberman Aiden, Jon Orwant, Will Brockman, and Slav Petrov. Syntactic annotations for the google books ngram corpus. In Proceedings of the ACL 2012 system demonstrations, pages 169–174. Association for Computational Linguistics, 2012.
  • Smith [Submitted] NJ Smith. ZS: A file format for efficiently distributing, using, and archiving record-oriented data sets of any size. Submitted.
  • Chen and Goodman [1999] SF Chen and J Goodman. An empirical study of smoothing techniques for language modeling. Comput Speech & Lang, 13(4):359–393, 1999.
  • Stolcke [2002] A Stolcke. Proceedings of icslp. volume 2, pages 901–904, 2002.