COMPS: Conceptual Minimal Pair Sentences for testing Property Knowledge and Inheritance in Pre-trained Language Models

10/05/2022
by   Kanishka Misra, et al.
9

A characteristic feature of human semantic memory is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) – i.e. demonstrate property inheritance. In this paper, we present COMPS, a collection of minimal pair sentences that jointly tests pre-trained language models (PLMs) on their ability to attribute properties to concepts and their ability to demonstrate property inheritance behavior. Analyses of 22 different PLMs on COMPS reveal that they can easily distinguish between concepts on the basis of a property when they are trivially different, but find it relatively difficult when concepts are related on the basis of nuanced knowledge representations. Furthermore, we find that PLMs can demonstrate behavior consistent with property inheritance to a great extent, but fail in the presence of distracting information, which decreases the performance of many models, sometimes even below chance. This lack of robustness in demonstrating simple reasoning raises important questions about PLMs' capacity to make correct inferences even when they appear to possess the prerequisite knowledge.

READ FULL TEXT

page 6

page 16

research
05/13/2022

A Property Induction Framework for Neural Language Models

To what extent can experience from language contribute to our conceptual...
research
05/22/2023

Can LLMs facilitate interpretation of pre-trained language models?

Work done to uncover the knowledge encoded within pre-trained language m...
research
10/06/2022

Modelling Commonsense Properties using Pre-Trained Bi-Encoders

Grasping the commonsense properties of everyday concepts is an important...
research
09/11/2022

Testing Pre-trained Language Models' Understanding of Distributivity via Causal Mediation Analysis

To what extent do pre-trained language models grasp semantic knowledge r...
research
02/28/2022

KMIR: A Benchmark for Evaluating Knowledge Memorization, Identification and Reasoning Abilities of Language Models

Previous works show the great potential of pre-trained language models (...
research
09/30/2021

Self-conditioning pre-trained language models

We study the presence of expert units in pre-trained Transformer-based L...
research
08/15/2022

Targeted Honeyword Generation with Language Models

Honeywords are fictitious passwords inserted into databases in order to ...

Please sign up or login with your details

Forgot password? Click here to reset