Compositional Demographic Word Embeddings

10/06/2020
by   Charles Welch, et al.
7

Word embeddings are usually derived from corpora containing text from many individuals, thus leading to general purpose representations rather than individually personalized representations. While personalized embeddings can be useful to improve language model performance and other language processing tasks, they can only be computed for people with a large amount of longitudinal data, which is not the case for new users. We propose a new form of personalized word embeddings that use demographic-specific word representations derived compositionally from full or partial demographic information for a user (i.e., gender, age, location, religion). We show that the resulting demographic-aware word representations outperform generic word representations on two tasks for English: language modeling and word associations. We further explore the trade-off between the number of available attributes and their relative effectiveness and discuss the ethical implications of using them.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/11/2020

Exploring the Value of Personalized Word Embeddings

In this paper, we introduce personalized word embeddings, and examine th...
research
04/08/2019

Evaluation of Greek Word Embeddings

Since word embeddings have been the most popular input for many NLP task...
research
04/30/2020

Analyzing the Surprising Variability in Word Embedding Stability Across Languages

Word embeddings are powerful representations that form the foundation of...
research
03/07/2017

Unsupervised Learning of Sentence Embeddings using Compositional n-Gram Features

The recent tremendous success of unsupervised word embeddings in a multi...
research
10/29/2017

Personalized word representations Carrying Personalized Semantics Learned from Social Network Posts

Distributed word representations have been shown to be very useful in va...
research
04/04/2022

"This is my unicorn, Fluffy": Personalizing frozen vision-language representations

Large Vision Language models pretrained on web-scale data provide re...
research
12/27/2019

Encoding word order in complex embeddings

Sequential word order is important when processing text. Currently, neur...

Please sign up or login with your details

Forgot password? Click here to reset