Injecting Knowledge into Biomedical Pre-trained Models via Polymorphism and Synonymous Substitution

05/24/2023
by   Hongbo Zhang, et al.
0

Pre-trained language models (PLMs) were considered to be able to store relational knowledge present in the training data. However, some relational knowledge seems to be discarded unsafely in PLMs due to report bias: low-frequency relational knowledge might be underexpressed compared to high-frequency one in PLMs. This gives us a hint that relational knowledge might not be redundant to the stored knowledge of PLMs, but rather be complementary. To additionally inject relational knowledge into PLMs, we propose a simple-yet-effective approach to inject relational knowledge into PLMs, which is inspired by three observations (namely, polymorphism, synonymous substitution, and association). In particular, we switch entities in the training corpus to related entities (either hypernyms/hyponyms/synonyms, or arbitrarily-related concepts). Experimental results show that the proposed approach could not only better capture relational knowledge, but also improve the performance in various biomedical downstream tasks. Our model is available in <https://github.com/StevenZHB/BioPLM_InjectingKnowledge>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/15/2023

Knowledge Rumination for Pre-trained Language Models

Previous studies have revealed that vanilla pre-trained language models ...
research
08/15/2023

A Comprehensive Study on Knowledge Graph Embedding over Relational Patterns Based on Rule Learning

Knowledge Graph Embedding (KGE) has proven to be an effective approach t...
research
05/03/2023

Causality-aware Concept Extraction based on Knowledge-guided Prompting

Concepts benefit natural language understanding but are far from complet...
research
04/12/2021

Relational world knowledge representation in contextual language models: A review

Relational knowledge bases (KBs) are established tools for world knowled...
research
08/28/2023

Biomedical Entity Linking with Triple-aware Pre-Training

Linking biomedical entities is an essential aspect in biomedical natural...
research
10/19/2022

Schema-aware Reference as Prompt Improves Data-Efficient Relational Triple and Event Extraction

Information Extraction, which aims to extract structural relational trip...
research
07/06/2023

Extracting Multi-valued Relations from Language Models

The widespread usage of latent language representations via pre-trained ...

Please sign up or login with your details

Forgot password? Click here to reset