KEPLET: Knowledge-Enhanced Pretrained Language Model with Topic Entity Awareness

05/02/2023
by   Yichuan Li, et al.
0

In recent years, Pre-trained Language Models (PLMs) have shown their superiority by pre-training on unstructured text corpus and then fine-tuning on downstream tasks. On entity-rich textual resources like Wikipedia, Knowledge-Enhanced PLMs (KEPLMs) incorporate the interactions between tokens and mentioned entities in pre-training, and are thus more effective on entity-centric tasks such as entity linking and relation classification. Although exploiting Wikipedia's rich structures to some extent, conventional KEPLMs still neglect a unique layout of the corpus where each Wikipedia page is around a topic entity (identified by the page URL and shown in the page title). In this paper, we demonstrate that KEPLMs without incorporating the topic entities will lead to insufficient entity interaction and biased (relation) word semantics. We thus propose KEPLET, a novel Knowledge-Enhanced Pre-trained LanguagE model with Topic entity awareness. In an end-to-end manner, KEPLET identifies where to add the topic entity's information in a Wikipedia sentence, fuses such information into token and mentioned entities representations, and supervises the network learning, through which it takes topic entities back into consideration. Experiments demonstrated the generality and superiority of KEPLET which was applied to two representative KEPLMs, achieving significant improvements on four entity-centric tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/20/2022

Representing Knowledge by Spans: A Knowledge-Enhanced Model for Information Extraction

Knowledge-enhanced pre-trained models for language representation have b...
research
08/31/2019

EntEval: A Holistic Evaluation Benchmark for Entity Representations

Rich entity representations are useful for a wide class of problems invo...
research
09/14/2023

Leveraging Contextual Information for Effective Entity Salience Detection

In text documents such as news articles, the content and key events usua...
research
05/22/2023

EnCore: Pre-Training Entity Encoders using Coreference Chains

Entity typing is the task of assigning semantic types to the entities th...
research
09/09/2019

Knowledge Enhanced Contextual Word Representations

Contextual word representations, typically trained on unstructured, unla...
research
10/21/2022

SpaBERT: A Pretrained Language Model from Geographic Data for Geo-Entity Representation

Named geographic entities (geo-entities for short) are the building bloc...
research
10/22/2022

Generative Prompt Tuning for Relation Classification

Using prompts to explore the knowledge contained within pre-trained lang...

Please sign up or login with your details

Forgot password? Click here to reset