Language coverage and generalization in RNN-based continuous sentence embeddings for interacting agents

11/05/2019
by   Luca Celotti, et al.
0

Continuous sentence embeddings using recurrent neural networks (RNNs), where variable-length sentences are encoded into fixed-dimensional vectors, are often the main building blocks of architectures applied to language tasks such as dialogue generation. While it is known that those embeddings are able to learn some structures of language (e.g. grammar) in a purely data-driven manner, there is very little work on the objective evaluation of their ability to cover the whole language space and to generalize to sentences outside the language bias of the training data. Using a manually designed context-free grammar (CFG) to generate a large-scale dataset of sentences related to the content of realistic 3D indoor scenes, we evaluate the language coverage and generalization abilities of the most common continuous sentence embeddings based on RNNs. We also propose a new embedding method based on arithmetic coding, AriEL, that is not data-driven and that efficiently encodes in continuous space any sentence from the CFG. We find that RNN-based embeddings underfit the training data and cover only a small subset of the language defined by the CFG. They also fail to learn the underlying CFG and generalize to unbiased sentences from that same CFG. We found that AriEL provides an insightful baseline.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/08/2019

In Search for Linear Relations in Sentence Embedding Spaces

We present an introductory investigation into continuous-space vector re...
research
08/15/2016

Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks

There is a lot of research interest in encoding variable length sentence...
research
02/11/2013

Toric grammars: a new statistical approach to natural language modeling

We propose a new statistical model for computational linguistics. Rather...
research
11/14/2018

CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling

In real-world applications of natural language generation, there are oft...
research
12/03/2019

COSTRA 1.0: A Dataset of Complex Sentence Transformations

We present COSTRA 1.0, a dataset of complex sentence transformations. Th...
research
05/25/2017

Neural Attribute Machines for Program Generation

Recurrent neural networks have achieved remarkable success at generating...

Please sign up or login with your details

Forgot password? Click here to reset