Decision-Directed Data Decomposition

09/18/2019
by   Brent D. Davis, et al.
0

We present an algorithm, Decision-Directed Data Decomposition, which decomposes a dataset into two components. The first contains most of the useful information for a specified supervised learning task, and the second orthogonal component that contains little information about the task. The algorithm is simple and scalable. It can use kernel techniques to help preserve desirable information in the decomposition. We illustrate its application to tasks in two domains, using distributed representations of words and images, and we report state-of-the-art results showcasing D_4's capability to remove information pertaining to gender from word embeddings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/29/2018

Learning Gender-Neutral Word Embeddings

Word embedding models have become a fundamental component in a wide rang...
research
12/15/2022

The effects of gender bias in word embeddings on depression prediction

Word embeddings are extensively used in various NLP problems as a state-...
research
08/05/2020

An exploration of the encoding of grammatical gender in word embeddings

The vector representation of words, known as word embeddings, has opened...
research
02/24/2016

Ultradense Word Embeddings by Orthogonal Transformation

Embeddings are generic representations that are useful for many NLP task...
research
10/30/2019

How does Grammatical Gender Affect Noun Representations in Gender-Marking Languages?

Many natural languages assign grammatical gender also to inanimate nouns...
research
02/07/2017

How to evaluate word embeddings? On importance of data efficiency and simple supervised tasks

Maybe the single most important goal of representation learning is makin...

Please sign up or login with your details

Forgot password? Click here to reset