Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference

09/08/2021
by   Xiaoyu Yang, et al.
0

While recent research on natural language inference has considerably benefited from large annotated datasets, the amount of inference-related knowledge (including commonsense) provided in the annotated data is still rather limited. There have been two lines of approaches that can be used to further address the limitation: (1) unsupervised pretraining can leverage knowledge in much larger unstructured text data; (2) structured (often human-curated) knowledge has started to be considered in neural-network-based models for NLI. An immediate question is whether these two approaches complement each other, or how to develop models that can bring together their advantages. In this paper, we propose models that leverage structured knowledge in different components of pre-trained models. Our results show that the proposed models perform better than previous BERT-based state-of-the-art models. Although our models are proposed for NLI, they can be easily extended to other sentence or sentence-pair classification problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2017

Natural Language Inference with External Knowledge

Modeling informal inference in natural language is very challenging. Wit...
research
08/19/2019

Align, Mask and Select: A Simple Method for Incorporating Commonsense Knowledge into Language Representation Models

Neural language representation models such as Bidirectional Encoder Repr...
research
12/08/2021

VIRT: Improving Representation-based Models for Text Matching through Virtual Interaction

With the booming of pre-trained transformers, remarkable progress has be...
research
04/03/2019

Unsupervised Deep Structured Semantic Models for Commonsense Reasoning

Commonsense reasoning is fundamental to natural language understanding. ...
research
11/03/2022

Eliciting Knowledge from Large Pre-Trained Models for Unsupervised Knowledge-Grounded Conversation

Recent advances in large-scale pre-training provide large models with th...
research
11/13/2019

Unsupervised Pre-training for Natural Language Generation: A Literature Review

Recently, unsupervised pre-training is gaining increasing popularity in ...
research
08/21/2018

Lessons from Natural Language Inference in the Clinical Domain

State of the art models using deep neural networks have become very good...

Please sign up or login with your details

Forgot password? Click here to reset