KI-BERT: Infusing Knowledge Context for Better Language and Domain Understanding

04/09/2021
by   Keyur Faldu, et al.
0

Contextualized entity representations learned by state-of-the-art deep learning models (BERT, GPT, T5, etc) leverage the attention mechanism to learn the data context. However, these models are still blind to leverage the knowledge context present in the knowledge graph. Knowledge context can be understood as semantics about entities, and their relationship with neighboring entities in knowledge graphs. We propose a novel and effective technique to infuse knowledge context from knowledge graphs for conceptual and ambiguous entities into models based on transformer architecture. Our novel technique project knowledge graph embedding in the homogeneous vector-space, introduces new token-types for entities, align entity position ids, and a selective attention mechanism. We take BERT as a baseline model and implement "KnowledgeInfused BERT" by infusing knowledge context from ConceptNet and WordNet, which significantly outperforms BERT over a wide range of NLP tasks over eight different GLUE datasets. KI-BERT-base model even outperforms BERT-large for domain-specific tasks like SciTail and academic subsets of QQP, QNLI, and MNLI.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/07/2019

KG-BERT: BERT for Knowledge Graph Completion

Knowledge graphs are important resources for many artificial intelligenc...
research
06/18/2022

Can Language Models Capture Graph Semantics? From Graphs to Language Model and Vice-Versa

Knowledge Graphs are a great resource to capture semantic knowledge in t...
research
03/01/2021

BERT-based knowledge extraction method of unstructured domain text

With the development and business adoption of knowledge graph, there is ...
research
09/21/2020

"When they say weed causes depression, but it's your fav antidepressant": Knowledge-aware Attention Framework for Relationship Extraction

With the increasing legalization of medical and recreational use of cann...
research
05/13/2018

An attention-based Bi-GRU-CapsNet model for hypernymy detection between compound entities

Named entities which composed of multiple continuous words frequently oc...
research
09/17/2019

K-BERT: Enabling Language Representation with Knowledge Graph

Pre-trained language representation models, such as BERT, capture a gene...
research
08/20/2019

Unsupervised Hierarchical Grouping of Knowledge Graph Entities

Knowledge graphs have attracted lots of attention in academic and indust...

Please sign up or login with your details

Forgot password? Click here to reset