Knowledge Enhanced Contextual Word Representations

09/09/2019
by   Matthew E. Peters, et al.
0

Contextual word representations, typically trained on unstructured, unlabeled text, do not contain any explicit grounding to real world entities and are often unable to remember facts about those entities. We propose a general method to embed multiple knowledge bases (KBs) into large scale models, and thereby enhance their representations with structured, human-curated knowledge. For each KB, we first use an integrated entity linker to retrieve relevant entity embeddings, then update contextual word representations via a form of word-to-entity attention. In contrast to previous approaches, the entity linkers and self-supervised language modeling objective are jointly trained end-to-end in a multitask setting that combines a small amount of entity linking supervision with a large amount of raw text. After integrating WordNet and a subset of Wikipedia into BERT, the knowledge enhanced BERT (KnowBert) demonstrates improved perplexity, ability to recall facts as measured in a probing task and downstream performance on relationship extraction, entity typing, and word sense disambiguation. KnowBert's runtime is comparable to BERT's and it scales to large KBs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2021

ExBERT: An External Knowledge Enhanced BERT for Natural Language Inference

Neural language representation models such as BERT, pre-trained on large...
research
01/11/2020

Learning Cross-Context Entity Representations from Text

Language modeling tasks, in which words, or word-pieces, are predicted o...
research
03/11/2020

Investigating Entity Knowledge in BERT with Simple Neural End-To-End Entity Linking

A typical architecture for end-to-end entity linking systems consists of...
research
05/02/2023

KEPLET: Knowledge-Enhanced Pretrained Language Model with Topic Entity Awareness

In recent years, Pre-trained Language Models (PLMs) have shown their sup...
research
03/31/2022

Scientific and Technological Text Knowledge Extraction Method of based on Word Mixing and GRU

The knowledge extraction task is to extract triple relations (head entit...
research
05/27/2023

Towards Better Entity Linking with Multi-View Enhanced Distillation

Dense retrieval is widely used for entity linking to retrieve entities f...
research
04/15/2020

Entities as Experts: Sparse Memory Access with Entity Supervision

We focus on the problem of capturing declarative knowledge in the learne...

Please sign up or login with your details

Forgot password? Click here to reset