Discovering Differences in the Representation of People using Contextualized Semantic Axes

10/21/2022
by   Li Lucy, et al.
0

A common paradigm for identifying semantic differences across social and temporal contexts is the use of static word embeddings and their distances. In particular, past work has compared embeddings against "semantic axes" that represent two opposing concepts. We extend this paradigm to BERT embeddings, and construct contextualized axes that mitigate the pitfall where antonyms have neighboring representations. We validate and demonstrate these axes on two people-centric datasets: occupations from Wikipedia, and multi-platform discussions in extremist, men's communities over fourteen years. In both studies, contextualized semantic axes can characterize differences among instances of the same word type. In the latter study, we show that references to women and the contexts around them have become more detestable over time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2017

Ontology-Aware Token Embeddings for Prepositional Phrase Attachment

Type-level word embeddings use the same set of parameters to represent a...
research
11/13/2019

What do you mean, BERT? Assessing BERT as a Distributional Semantics Model

Contextualized word embeddings, i.e. vector representations for words in...
research
12/30/2020

SemGloVe: Semantic Co-occurrences for GloVe from BERT

GloVe learns word embeddings by leveraging statistical information from ...
research
08/30/2019

Automatically Inferring Gender Associations from Language

In this paper, we pose the question: do people talk about women and men ...
research
02/11/2023

Dialectograms: Machine Learning Differences between Discursive Communities

Word embeddings provide an unsupervised way to understand differences in...
research
05/16/2023

Measuring Stereotypes using Entity-Centric Data

Stereotypes inform how we present ourselves and others, and in turn how ...
research
11/13/2020

Learning language variations in news corpora through differential embeddings

There is an increasing interest in the NLP community in capturing variat...

Please sign up or login with your details

Forgot password? Click here to reset