Correcting the Common Discourse Bias in Linear Representation of Sentences using Conceptors

11/17/2018
by   Tianlin Liu, et al.
0

Distributed representations of words, better known as word embeddings, have become important building blocks for natural language processing tasks. Numerous studies are devoted to transferring the success of unsupervised word embeddings to sentence embeddings. In this paper, we introduce a simple representation of sentences in which a sentence embedding is represented as a weighted average of word vectors followed by a soft projection. We demonstrate the effectiveness of this proposed method on the clinical semantic textual similarity task of the BioCreative/OHNLP Challenge 2018.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2017

Unsupervised Sentence Representations as Word Information Series: Revisiting TF--IDF

Sentence representation at the semantic level is a challenging task for ...
research
03/07/2017

Unsupervised Learning of Sentence Embeddings using Compositional n-Gram Features

The recent tremendous success of unsupervised word embeddings in a multi...
research
10/23/2020

Domain Specific Complex Sentence (DCSC) Semantic Similarity Dataset

Semantic textual similarity is one of the open research challenges in th...
research
11/11/2022

Improving word mover's distance by leveraging self-attention matrix

Measuring the semantic similarity between two sentences is still an impo...
research
04/18/2017

Representing Sentences as Low-Rank Subspaces

Sentences are important semantic units of natural language. A generic, d...
research
05/16/2020

Learning Probabilistic Sentence Representations from Paraphrases

Probabilistic word embeddings have shown effectiveness in capturing noti...
research
05/10/2021

DefSent: Sentence Embeddings using Definition Sentences

Sentence embedding methods using natural language inference (NLI) datase...

Please sign up or login with your details

Forgot password? Click here to reset