The paradox of the compositionality of natural language: a neural machine translation case study

08/12/2021
by   Verna Dankers, et al.
0

Moving towards human-like linguistic performance is often argued to require compositional generalisation. Whether neural networks exhibit this ability is typically studied using artificial languages, for which the compositionality of input fragments can be guaranteed and their meanings algebraically composed. However, compositionality in natural language is vastly more complex than this rigid, arithmetics-like version of compositionality, and as such artificial compositionality tests do not allow us to draw conclusions about how neural models deal with compositionality in more realistic scenarios. In this work, we re-instantiate three compositionality tests from the literature and reformulate them for neural machine translation (NMT). The results highlight two main issues: the inconsistent behaviour of NMT models and their inability to (correctly) modulate between local and global processing. Aside from an empirical study, our work is a call to action: we should rethink the evaluation of compositionality in neural networks of natural language, where composing meaning is not as straightforward as doing the math.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2023

Neural Machine Translation for Code Generation

Neural machine translation (NMT) methods developed for natural language ...
research
12/18/2019

A Survey on Document-level Machine Translation: Methods and Evaluation

Machine translation (MT) is an important task in natural language proces...
research
08/15/2023

Anaphoric Structure Emerges Between Neural Networks

Pragmatics is core to natural language, enabling speakers to communicate...
research
10/02/2022

The boundaries of meaning: a case study in neural machine translation

The success of deep learning in natural language processing raises intri...
research
02/20/2021

Understanding and Enhancing the Use of Context for Machine Translation

To understand and infer meaning in language, neural models have to learn...
research
01/31/2023

Recursive Neural Networks with Bottlenecks Diagnose (Non-)Compositionality

A recent line of work in NLP focuses on the (dis)ability of models to ge...
research
02/08/2022

Can We Generate Shellcodes via Natural Language? An Empirical Study

Writing software exploits is an important practice for offensive securit...

Please sign up or login with your details

Forgot password? Click here to reset