Efficient and Interpretable Neural Models for Entity Tracking

08/30/2022
by   Shubham Toshniwal, et al.
6

What would it take for a natural language model to understand a novel, such as The Lord of the Rings? Among other things, such a model must be able to: (a) identify and record new characters (entities) and their attributes as they are introduced in the text, and (b) identify subsequent references to the characters previously introduced and update their attributes. This problem of entity tracking is essential for language understanding, and thus, useful for a wide array of downstream applications in NLP such as question-answering, summarization. In this thesis, we focus on two key problems in relation to facilitating the use of entity tracking models: (i) scaling entity tracking models to long documents, such as a novel, and (ii) integrating entity tracking into language models. Applying language technologies to long documents has garnered interest recently, but computational constraints are a significant bottleneck in scaling up current methods. In this thesis, we argue that computationally efficient entity tracking models can be developed by representing entities with rich, fixed-dimensional vector representations derived from pretrained language models, and by exploiting the ephemeral nature of entities. We also argue for the integration of entity tracking into language models as it will allow for: (i) wider application given the current ubiquitous use of pretrained language models in NLP applications, and (ii) easier adoption since it is much easier to swap in a new pretrained language model than to integrate a separate standalone entity tracking model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/02/2017

Dynamic Entity Representations in Neural Language Models

Understanding a long document requires tracking how entities are introdu...
research
05/03/2023

Entity Tracking in Language Models

Keeping track of how states and relations of entities change as a text o...
research
12/18/2022

Can Retriever-Augmented Language Models Reason? The Blame Game Between the Retriever and the Language Model

The emergence of large pretrained models has enabled language models to ...
research
10/21/2022

SpaBERT: A Pretrained Language Model from Geographic Data for Geo-Entity Representation

Named geographic entities (geo-entities for short) are the building bloc...
research
12/31/2020

On the importance of functions in data modeling

In this paper we argue that representing entity properties by tuple attr...
research
11/04/2021

CoreLM: Coreference-aware Language Model Fine-Tuning

Language Models are the underpin of all modern Natural Language Processi...
research
12/31/2020

Understanding Politics via Contextualized Discourse Processing

Politicians often have underlying agendas when reacting to events. Argum...

Please sign up or login with your details

Forgot password? Click here to reset