Toric grammars: a new statistical approach to natural language modeling

02/11/2013
by   Olivier Catoni, et al.
0

We propose a new statistical model for computational linguistics. Rather than trying to estimate directly the probability distribution of a random sentence of the language, we define a Markov chain on finite sets of sentences with many finite recurrent communicating classes and define our language model as the invariant probability measures of the chain on each recurrent communicating class. This Markov chain, that we call a communication model, recombines at each step randomly the set of sentences forming its current state, using some grammar rules. When the grammar rules are fixed and known in advance instead of being estimated on the fly, we can prove supplementary mathematical properties. In particular, we can prove in this case that all states are recurrent states, so that the chain defines a partition of its state space into finite recurrent communicating classes. We show that our approach is a decisive departure from Markov models at the sentence level and discuss its relationships with Context Free Grammars. Although the toric grammars we use are closely related to Context Free Grammars, the way we generate the language from the grammar is qualitatively different. Our communication model has two purposes. On the one hand, it is used to define indirectly the probability distribution of a random sentence of the language. On the other hand it can serve as a (crude) model of language transmission from one speaker to another speaker through the communication of a (large) set of sentences.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2022

Learning grammar with a divide-and-concur neural network

We implement a divide-and-concur iterative projection approach to contex...
research
11/05/2019

Language coverage and generalization in RNN-based continuous sentence embeddings for interacting agents

Continuous sentence embeddings using recurrent neural networks (RNNs), w...
research
03/14/2021

Learning a Word-Level Language Model with Sentence-Level Noise Contrastive Estimation for Contextual Sentence Probability Estimation

Inferring the probability distribution of sentences or word sequences is...
research
06/26/2020

Dissecting Power of a Finite Intersection of Context Free Languages

Let ^k,α denote a tetration function defined as follows: ^1,α=2^α and ^k...
research
04/01/2016

A Compositional Approach to Language Modeling

Traditional language models treat language as a finite state automaton o...
research
09/04/2018

Random Language Model: a path to principled complexity

Many complex generative systems use languages to create structured objec...
research
06/14/2021

Grammar Equations

Diagrammatically speaking, grammatical calculi such as pregroups provide...

Please sign up or login with your details

Forgot password? Click here to reset