Why So Down? The Role of Negative (and Positive) Pointwise Mutual Information in Distributional Semantics

08/19/2019
by   Alexandre Salle, et al.
0

In distributional semantics, the pointwise mutual information (PMI) weighting of the cooccurrence matrix performs far better than raw counts. There is, however, an issue with unobserved pair cooccurrences as PMI goes to negative infinity. This problem is aggravated by unreliable statistics from finite corpora which lead to a large number of such pairs. A common practice is to clip negative PMI (- PMI) at 0, also known as Positive PMI (PPMI). In this paper, we investigate alternative ways of dealing with - PMI and, more importantly, study the role that negative information plays in the performance of a low-rank, weighted factorization of different PMI matrices. Using various semantic and syntactic tasks as probes into models which use either negative or positive PMI (or both), we find that most of the encoded semantics and syntax come from positive PMI, in contrast to - PMI which contributes almost exclusively syntactic information. Our findings deepen our understanding of distributional semantics, while also introducing novel PMI variants and grounding the popular PPMI measure.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2016

Matrix Factorization using Window Sampling and Negative Sampling for Improved Word Representations

In this paper, we propose LexVec, a new method for generating distribute...
research
07/24/2019

Distributional Analysis of Function Words

This paper is a first attempt at reconciling the current methods of dist...
research
09/30/2022

Synonym Detection Using Syntactic Dependency And Neural Embeddings

Recent advances on the Vector Space Model have significantly improved so...
research
05/27/2021

Rethinking InfoNCE: How Many Negative Samples Do You Need?

InfoNCE loss is a widely used loss function for contrastive model traini...
research
06/03/2016

Enhancing the LexVec Distributed Word Representation Model Using Positional Contexts and External Memory

In this paper we take a state-of-the-art model for distributed word repr...
research
12/01/2018

Learning Speaker Representations with Mutual Information

Learning good representations is of crucial importance in deep learning....
research
12/22/2015

Restricted Predicates for Hypothetical Datalog

Hypothetical Datalog is based on an intuitionistic semantics rather than...

Please sign up or login with your details

Forgot password? Click here to reset