Improving word mover's distance by leveraging self-attention matrix

11/11/2022
by   Hiroaki Yamagiwa, et al.
0

Measuring the semantic similarity between two sentences is still an important task. The word mover's distance (WMD) computes the similarity via the optimal alignment between the sets of word embeddings. However, WMD does not utilize word order, making it difficult to distinguish sentences with large overlaps of similar words, even if they are semantically very different. Here, we attempt to improve WMD by incorporating the sentence structure represented by BERT's self-attention matrix (SAM). The proposed method is based on the Fused Gromov-Wasserstein distance, which simultaneously considers the similarity of the word embedding and the SAM for calculating the optimal transport between two sentences. Experiments on paraphrase identification and semantic textual similarity show that the proposed method improves WMD and its variants. Our code is available at https://github.com/ymgw55/WSMD.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/17/2018

Correcting the Common Discourse Bias in Linear Representation of Sentences using Conceptors

Distributed representations of words, better known as word embeddings, h...
research
06/07/2023

Unbalanced Optimal Transport for Unbalanced Word Alignment

Monolingual word alignment is crucial to model semantic interactions bet...
research
07/15/2020

Logic Constrained Pointer Networks for Interpretable Textual Similarity

Systematically discovering semantic relationships in text is an importan...
research
04/11/2019

Gating Mechanisms for Combining Character and Word-level Word Representations: An Empirical Study

In this paper we study how different ways of combining character and wor...
research
09/12/2018

Distilled Wasserstein Learning for Word Embedding and Topic Modeling

We propose a novel Wasserstein method with a distillation mechanism, yie...
research
11/07/2019

Explicit Pairwise Word Interaction Modeling Improves Pretrained Transformers for English Semantic Similarity Tasks

In English semantic similarity tasks, classic word embedding-based appro...
research
07/22/2022

Exploring Wasserstein Distance across Concept Embeddings for Ontology Matching

Measuring the distance between ontological elements is a fundamental com...

Please sign up or login with your details

Forgot password? Click here to reset