Semantic sentence similarity: size does not always matter

06/16/2021
by   Danny Merkx, et al.
0

This study addresses the question whether visually grounded speech recognition (VGS) models learn to capture sentence semantics without access to any prior linguistic knowledge. We produce synthetic and natural spoken versions of a well known semantic textual similarity database and show that our VGS model produces embeddings that correlate well with human semantic similarity judgements. Our results show that a model trained on a small image-caption database outperforms two models trained on much larger databases, indicating that database size is not all that matters. We also investigate the importance of having multiple captions per image and find that this is indeed helpful even if the total number of images is lower, suggesting that paraphrasing is a valuable learning signal. While the general trend in the field is to create ever larger datasets to train models on, our findings indicate other characteristics of the database can just as important important.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/27/2019

Learning semantic sentence representations from visually grounded language without lexical knowledge

Current approaches to learning semantic representations of sentences oft...
research
04/13/2015

Joint Learning of Distributed Representations for Images and Texts

This technical report provides extra details of the deep multimodal simi...
research
03/14/2022

Modelling word learning and recognition using visually grounded speech

Background: Computational models of speech recognition often assume that...
research
09/09/2019

Language learning using Speech to Image retrieval

Humans learn language by interaction with their environment and listenin...
research
02/09/2011

Ologs: a categorical framework for knowledge representation

In this paper we introduce the olog, or ontology log, a category-theoret...
research
02/07/2017

Representations of language in a model of visually grounded speech signal

We present a visually grounded model of speech perception which projects...
research
04/20/2018

Sequential Network Transfer: Adapting Sentence Embeddings to Human Activities and Beyond

We study the problem of adapting neural sentence embedding models to the...

Please sign up or login with your details

Forgot password? Click here to reset