Sliced Mutual Information: A Scalable Measure of Statistical Dependence

10/11/2021
by   Ziv Goldfeld, et al.
0

Mutual information (MI) is a fundamental measure of statistical dependence, with a myriad of applications to information theory, statistics, and machine learning. While it possesses many desirable structural properties, the estimation of high-dimensional MI from samples suffers from the curse of dimensionality. Motivated by statistical scalability to high dimensions, this paper proposes sliced MI (SMI) as a surrogate measure of dependence. SMI is defined as an average of MI terms between one-dimensional random projections. We show that it preserves many of the structural properties of classic MI, while gaining scalable computation and efficient estimation from samples. Furthermore, and in contrast to classic MI, SMI can grow as a result of deterministic transformations. This enables leveraging SMI for feature extraction by optimizing it over processing functions of raw data to identify useful representations thereof. Our theory is supported by numerical studies of independence testing and feature extraction, which demonstrate the potential gains SMI offers over classic MI for high-dimensional inference.

READ FULL TEXT

page 8

page 11

research
06/17/2022

k-Sliced Mutual Information: A Quantitative Study of Scalability with Dimension

Sliced mutual information (SMI) is defined as an average of mutual infor...
research
08/16/2022

Measuring Statistical Dependencies via Maximum Norm and Characteristic Functions

In this paper, we focus on the problem of statistical dependence estimat...
research
07/14/2012

Dimension Reduction by Mutual Information Feature Extraction

During the past decades, to study high-dimensional data in a large varie...
research
06/10/2012

Dimension Reduction by Mutual Information Discriminant Analysis

In the past few decades, researchers have proposed many discriminant ana...
research
12/01/2018

Learning Speaker Representations with Mutual Information

Learning good representations is of crucial importance in deep learning....
research
06/16/2016

Estimating mutual information in high dimensions via classification error

Multivariate pattern analyses approaches in neuroimaging are fundamental...
research
05/09/2015

Measuring dependence powerfully and equitably

Given a high-dimensional data set we often wish to find the strongest re...

Please sign up or login with your details

Forgot password? Click here to reset