Exploring Transitivity in Neural NLI Models through Veridicality

01/26/2021
by   Hitomi Yanaka, et al.
0

Despite the recent success of deep neural networks in natural language processing, the extent to which they can demonstrate human-like generalization capacities for natural language understanding remains unclear. We explore this issue in the domain of natural language inference (NLI), focusing on the transitivity of inference relations, a fundamental property for systematically drawing inferences. A model capturing transitivity can compose basic inference patterns and draw new inferences. We introduce an analysis method using synthetic and naturalistic NLI datasets involving clause-embedding verbs to evaluate whether models can perform transitivity inferences composed of veridical inferences and arbitrary inference types. We find that current NLI models do not perform consistently well on transitivity inference tasks, suggesting that they lack the generalization capacity for drawing composite inferences from provided training examples. The data and code for our analysis are publicly available at https://github.com/verypluming/transitivity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/30/2020

Do Neural Models Learn Systematicity of Monotonicity Inference in Natural Language?

Despite the success of language models using neural networks, it remains...
research
04/07/2020

Are Natural Language Inference Models IMPPRESsive? Learning IMPlicature and PRESupposition

Natural language inference (NLI) is an increasingly important task for n...
research
12/14/2021

Representing Inferences and their Lexicalization

We have recently begun a project to develop a more effective and efficie...
research
10/20/2020

ConjNLI: Natural Language Inference Over Conjunctive Sentences

Reasoning about conjuncts in conjunctive sentences is important for a de...
research
09/06/2019

Uncertain Natural Language Inference

We propose a refinement of Natural Language Inference (NLI), called Unce...
research
10/31/2019

Harnessing the richness of the linguistic signal in predicting pragmatic inferences

The strength of pragmatic inferences systematically depends on linguisti...
research
09/14/2022

Drawing Causal Inferences About Performance Effects in NLP

This article emphasizes that NLP as a science seeks to make inferences a...

Please sign up or login with your details

Forgot password? Click here to reset