Representing Knowledge by Spans: A Knowledge-Enhanced Model for Information Extraction

08/20/2022
by   Jiacheng Li, et al.
0

Knowledge-enhanced pre-trained models for language representation have been shown to be more effective in knowledge base construction tasks (i.e., relation extraction) than language models such as BERT. These knowledge-enhanced language models incorporate knowledge into pre-training to generate representations of entities or relationships. However, existing methods typically represent each entity with a separate embedding. As a result, these methods struggle to represent out-of-vocabulary entities and a large amount of parameters, on top of their underlying token models (i.e., the transformer), must be used and the number of entities that can be handled is limited in practice due to memory constraints. Moreover, existing models still struggle to represent entities and relationships simultaneously. To address these problems, we propose a new pre-trained model that learns representations of both entities and relationships from token spans and span pairs in the text respectively. By encoding spans efficiently with span modules, our model can represent both entities and their relationships but requires fewer parameters than existing models. We pre-trained our model with the knowledge graph extracted from Wikipedia and test it on a broad range of supervised and unsupervised information extraction tasks. Results show that our model learns better representations for both entities and relationships than baselines, while in supervised settings, fine-tuning our model outperforms RoBERTa consistently and achieves competitive results on information extraction tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/13/2019

KEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation

Pre-trained language representation models (PLMs) learn effective langua...
research
05/02/2023

KEPLET: Knowledge-Enhanced Pretrained Language Model with Topic Entity Awareness

In recent years, Pre-trained Language Models (PLMs) have shown their sup...
research
12/02/2021

DKPLM: Decomposable Knowledge-enhanced Pre-trained Language Model for Natural Language Understanding

Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained ...
research
01/27/2022

Ontology-enhanced Prompt-tuning for Few-shot Learning

Few-shot Learning (FSL) is aimed to make predictions based on a limited ...
research
12/15/2021

GenIE: Generative Information Extraction

Structured and grounded representation of text is typically formalized b...
research
01/26/2023

Understanding Finetuning for Factual Knowledge Extraction from Language Models

Language models (LMs) pretrained on large corpora of text from the web h...
research
09/10/2020

RadLex Normalization in Radiology Reports

Radiology reports have been widely used for extraction of various clinic...

Please sign up or login with your details

Forgot password? Click here to reset