Entities as Experts: Sparse Memory Access with Entity Supervision

04/15/2020
by   Thibault Févry, et al.
0

We focus on the problem of capturing declarative knowledge in the learned parameters of a language model. We introduce a new model, Entities as Experts (EaE), that can access distinct memories of the entities mentioned in a piece of text. Unlike previous efforts to integrate entity knowledge into sequence models, EaE's entity representations are learned directly from text. These representations capture sufficient knowledge to answer TriviaQA questions such as "Which Dr. Who villain has been played by Roger Delgado, Anthony Ainley, Eric Roberts?". EaE outperforms a Transformer model with 30× the parameters on this task. According to the Lama knowledge probes, EaE also contains more factual knowledge than a similar sized Bert. We show that associating parameters with specific entities means that EaE only needs to access a fraction of its parameters at inference time, and we show that the correct identification, and representation, of entities is essential to EaE's performance. We also argue that the discrete and independent entity representations in EaE make it more modular and interpretable than the Transformer architecture on which it is based.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/11/2020

Learning Cross-Context Entity Representations from Text

Language modeling tasks, in which words, or word-pieces, are predicted o...
research
04/09/2019

Knowledge-Augmented Language Model and its Application to Unsupervised Named-Entity Recognition

Traditional language models are unable to efficiently model entity names...
research
10/02/2020

LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention

Entity representations are useful in natural language tasks involving en...
research
09/13/2022

Entity Tagging: Extracting Entities in Text Without Mention Supervision

Detection and disambiguation of all entities in text is a crucial task f...
research
10/12/2021

Mention Memory: incorporating textual knowledge into Transformers through entity mention attention

Natural language understanding tasks such as open-domain question answer...
research
09/09/2019

Knowledge Enhanced Contextual Word Representations

Contextual word representations, typically trained on unstructured, unla...
research
07/25/2017

Structural Regularities in Text-based Entity Vector Spaces

Entity retrieval is the task of finding entities such as people or produ...

Please sign up or login with your details

Forgot password? Click here to reset