ERICA: Improving Entity and Relation Understanding for Pre-trained Language Models via Contrastive Learning

12/30/2020
by   Yujia Qin, et al.
17

Pre-trained Language Models (PLMs) have shown strong performance in various downstream Natural Language Processing (NLP) tasks. However, PLMs still cannot well capture the factual knowledge in the text, which is crucial for understanding the whole text, especially for document-level language understanding tasks. To address this issue, we propose a novel contrastive learning framework named ERICA in pre-training phase to obtain a deeper understanding of the entities and their relations in text. Specifically, (1) to better understand entities, we propose an entity discrimination task that distinguishes which tail entity can be inferred by the given head entity and relation. (2) Besides, to better understand relations, we employ a relation discrimination task which distinguishes whether two entity pairs are close or not in relational semantics. Experimental results demonstrate that our proposed ERICA framework achieves consistent improvements on several document-level language understanding tasks, including relation extraction and reading comprehension, especially under low resource setting. Meanwhile, ERICA achieves comparable or better performance on sentence-level tasks. We will release the datasets, source codes and pre-trained language models for further research explorations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/01/2022

DictBERT: Dictionary Description Knowledge Enhanced Language Model Pre-training via Contrastive Learning

Although pre-trained language models (PLMs) have achieved state-of-the-a...
research
10/11/2022

Revisiting and Advancing Chinese Natural Language Understanding with Accelerated Heterogeneous Knowledge Pre-training

Recently, knowledge-enhanced pre-trained language models (KEPLMs) improv...
research
05/02/2023

UNTER: A Unified Knowledge Interface for Enhancing Pre-trained Language Models

Recent research demonstrates that external knowledge injection can advan...
research
05/05/2022

Relational Representation Learning in Visually-Rich Documents

Relational understanding is critical for a number of visually-rich docum...
research
11/20/2022

Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge

Contextual synonym knowledge is crucial for those similarity-oriented ta...
research
05/08/2023

Language Independent Neuro-Symbolic Semantic Parsing for Form Understanding

Recent works on form understanding mostly employ multimodal transformers...
research
11/29/2021

SimCLAD: A Simple Framework for Contrastive Learning of Acronym Disambiguation

Acronym disambiguation means finding the correct meaning of an ambiguous...

Please sign up or login with your details

Forgot password? Click here to reset