Are "Undocumented Workers" the Same as "Illegal Aliens"? Disentangling Denotation and Connotation in Vector Spaces

10/06/2020
by   Albert Webson, et al.
0

In politics, neologisms are frequently invented for partisan objectives. For example, "undocumented workers" and "illegal aliens" refer to the same group of people (i.e., they have the same denotation), but they carry clearly different connotations. Examples like these have traditionally posed a challenge to reference-based semantic theories and led to increasing acceptance of alternative theories (e.g., Two-Factor Semantics) among philosophers and cognitive scientists. In NLP, however, popular pretrained models encode both denotation and connotation as one entangled representation. In this study, we propose an adversarial nerual netowrk that decomposes a pretrained representation as independent denotation and connotation representations. For intrinsic interpretability, we show that words with the same denotation but different connotations (e.g., "immigrants" vs. "aliens", "estate tax" vs. "death tax") move closer to each other in denotation space while moving further apart in connotation space. For extrinsic application, we train an information retrieval system with our disentangled representations and show that the denotation vectors improve the viewpoint diversity of document rankings.

READ FULL TEXT
research
04/14/2021

Disentangling Representations of Text by Masking Transformers

Representations from large pretrained models such as BERT encode a range...
research
06/28/2021

Word2Box: Learning Word Representation Using Box Embeddings

Learning vector representations for words is one of the most fundamental...
research
11/16/2020

Comparative Probing of Lexical Semantics Theories for Cognitive Plausibility and Technological Usefulness

Lexical semantics theories differ in advocating that the meaning of word...
research
02/12/2023

Policy-Induced Self-Supervision Improves Representation Finetuning in Visual RL

We study how to transfer representations pretrained on source tasks to t...
research
04/19/2018

Learning Disentangled Representations of Texts with Application to Biomedical Abstracts

We propose a method for learning disentangled sets of vector representat...
research
06/22/2023

Identifying and Disentangling Spurious Features in Pretrained Image Representations

Neural networks employ spurious correlations in their predictions, resul...
research
04/18/2022

Active Learning Helps Pretrained Models Learn the Intended Task

Models can fail in unpredictable ways during deployment due to task ambi...

Please sign up or login with your details

Forgot password? Click here to reset