Modelling Compositionality and Structure Dependence in Natural Language

Human beings possess the most sophisticated computational machinery in the known universe. We can understand language of rich descriptive power, and communicate in the same environment with astonishing clarity. Two of the many contributors to the interest in natural language - the properties of Compositionality and Structure Dependence, are well documented, and offer a vast space to ask interesting modelling questions. The first step to begin answering these questions is to ground verbal theory in formal terms. Drawing on linguistics and set theory, a formalisation of these ideas is presented in the first half of this thesis. We see how cognitive systems that process language need to have certain functional constraints, viz. time based, incremental operations that rely on a structurally defined domain. The observations that result from analysing this formal setup are examined as part of a modelling exercise. Using the advances of word embedding techniques, a model of relational learning is simulated with a custom dataset to demonstrate how a time based role-filler binding mechanism satisfies some of the constraints described in the first section. The model's ability to map structure, along with its symbolic-connectionist architecture makes for a cognitively plausible implementation. The formalisation and simulation are together an attempt to recognise the constraints imposed by linguistic theory, and explore the opportunities presented by a cognitive model of relation learning to realise these constraints.

READ FULL TEXT
research
03/13/2017

El Lenguaje Natural como Lenguaje Formal

Formal languages theory is useful for the study of natural language. In ...
research
02/22/2016

From quantum foundations via natural language meaning to a theory of everything

In this paper we argue for a paradigmatic shift from `reductionism' to `...
research
07/17/2020

Toward Givenness Hierarchy Theoretic Natural Language Generation

Language-capable interactive robots participating in dialogues with huma...
research
10/05/2019

Natural- to formal-language generation using Tensor Product Representations

Generating formal-language represented by relational tuples, such as Lis...
research
12/14/2022

Intensional First Order Logic for Strong-AI Generation of Robots

Neuro-symbolic AI attempts to integrate neural and symbolic architecture...
research
07/01/2016

Throwing fuel on the embers: Probability or Dichotomy, Cognitive or Linguistic?

Prof. Robert Berwick's abstract for his forthcoming invited talk at the ...
research
12/13/2004

Vector Symbolic Architectures answer Jackendoff's challenges for cognitive neuroscience

Jackendoff (2002) posed four challenges that linguistic combinatoriality...

Please sign up or login with your details

Forgot password? Click here to reset