Recursive Neural Networks with Bottlenecks Diagnose (Non-)Compositionality

01/31/2023
by   Verna Dankers, et al.
0

A recent line of work in NLP focuses on the (dis)ability of models to generalise compositionally for artificial languages. However, when considering natural language tasks, the data involved is not strictly, or locally, compositional. Quantifying the compositionality of data is a challenging task, which has been investigated primarily for short utterances. We use recursive neural models (Tree-LSTMs) with bottlenecks that limit the transfer of information between nodes. We illustrate that comparing data's representations in models with and without the bottleneck can be used to produce a compositionality metric. The procedure is applied to the evaluation of arithmetic expressions using synthetic data, and sentiment classification using natural language data. We demonstrate that compression through a bottleneck impacts non-compositional examples disproportionately and then use the bottleneck compositionality metric (BCM) to distinguish compositional from non-compositional samples, yielding a compositionality ranking over a dataset.

READ FULL TEXT
research
06/19/2021

Improving Compositional Generalization in Classification Tasks via Structure Annotations

Compositional generalization is the ability to generalize systematically...
research
04/28/2020

Unnatural Language Processing: Bridging the Gap Between Synthetic and Natural Language Data

Large, human-annotated datasets are central to the development of natura...
research
05/11/2017

Dynamic Compositional Neural Networks over Tree Structure

Tree-structured neural networks have proven to be effective in learning ...
research
08/12/2021

The paradox of the compositionality of natural language: a neural machine translation case study

Moving towards human-like linguistic performance is often argued to requ...
research
04/27/2020

Word Interdependence Exposes How LSTMs Compose Representations

Recent work in NLP shows that LSTM language models capture compositional...
research
07/08/2022

Getting BART to Ride the Idiomatic Train: Learning to Represent Idiomatic Expressions

Idiomatic expressions (IEs), characterized by their non-compositionality...
research
03/10/2021

Relational Weight Priors in Neural Networks for Abstract Pattern Learning and Language Modelling

Deep neural networks have become the dominant approach in natural langua...

Please sign up or login with your details

Forgot password? Click here to reset