K-AID: Enhancing Pre-trained Language Models with Domain Knowledge for Question Answering

09/22/2021
by   Fu Sun, et al.
0

Knowledge enhanced pre-trained language models (K-PLMs) are shown to be effective for many public tasks in the literature but few of them have been successfully applied in practice. To address this problem, we propose K-AID, a systematic approach that includes a low-cost knowledge acquisition process for acquiring domain knowledge, an effective knowledge infusion module for improving model performance, and a knowledge distillation component for reducing the model size and deploying K-PLMs on resource-restricted devices (e.g., CPU) for real-world application. Importantly, instead of capturing entity knowledge like the majority of existing K-PLMs, our approach captures relational knowledge, which contributes to better-improving sentence-level text classification and text matching tasks that play a key role in question answering (QA). We conducted a set of experiments on five text classification tasks and three text matching tasks from three domains, namely E-commerce, Government, and Film TV, and performed online A/B tests in E-commerce. Experimental results show that our approach is able to achieve substantial improvement on sentence-level question answering tasks and bring beneficial business value in industrial settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/24/2020

AliMe KG: Domain Knowledge Graph Construction and Application in E-commerce

Pre-sales customer service is of importance to E-commerce platforms as i...
research
04/19/2019

Unifying Question Answering and Text Classification via Span Extraction

Even as pre-trained language encoders such as BERT are shared across man...
research
08/19/2019

Question Answering based Clinical Text Structuring Using Pre-trained Language Model

Clinical text structuring is a critical and fundamental task for clinica...
research
10/07/2020

DiPair: Fast and Accurate Distillation for Trillion-Scale Text Matching and Pair Modeling

Pre-trained models like BERT (Devlin et al., 2018) have dominated NLP / ...
research
07/11/2022

Embedding Recycling for Language Models

Training and inference with large neural models is expensive. However, f...
research
12/12/2022

Improving Generalization of Pre-trained Language Models via Stochastic Weight Averaging

Knowledge Distillation (KD) is a commonly used technique for improving t...
research
12/12/2019

AliMe KBQA: Question Answering over Structured Knowledge for E-commerce Customer Service

With the rise of knowledge graph (KG), question answering over knowledge...

Please sign up or login with your details

Forgot password? Click here to reset