Distance-Based Independence Screening for Canonical Analysis

02/28/2019
by   Chuanping Yu, et al.
0

This paper introduces a new method named Distance-based Independence Screening for Canonical Analysis (DISCA) to reduce dimensions of two random vectors with arbitrary dimensions. The objective of our method is to identify the low dimensional linear projections of two random vectors, such that any dimension reduction based on linear projection with lower dimensions will surely affect some dependent structure -- the removed components are not independent. The essence of DISCA is to use the distance correlation to eliminate the "redundant" dimensions until infeasible. Unlike the existing canonical analysis methods, DISCA does not require the dimensions of the reduced subspaces of the two random vectors to be equal, nor does it require certain distributional assumption on the random vectors. We show that under mild conditions, our approach does undercover the lowest possible linear dependency structures between two random vectors, and our conditions are weaker than some sufficient linear subspace-based methods. Numerically, DISCA is to solve a non-convex optimization problem. We formulate it as a difference-of-convex (DC) optimization problem, and then further adopt the alternating direction method of multipliers (ADMM) on the convex step of the DC algorithms to parallelize/accelerate the computation. Some sufficient linear subspace-based methods use potentially numerically-intensive bootstrap method to determine the dimensions of the reduced subspaces in advance; our method avoids this complexity. In simulations, we present cases that DISCA can solve effectively, while other methods cannot. In both the simulation studies and real data cases, when the other state-of-the-art dimension reduction methods are applicable, we observe that DISCA performs either comparably or better than most of them. Codes and an R package can be found in GitHub https://github.com/ChuanpingYu/DISCA.

READ FULL TEXT
research
05/08/2018

Optimal Subspace Estimation Using Overidentifying Vectors via Generalized Method of Moments

Many statistical models seek relationship between variables via subspace...
research
12/13/2019

MM Algorithms for Distance Covariance based Sufficient Dimension Reduction and Sufficient Variable Selection

Sufficient dimension reduction (SDR) using distance covariance (DCOV) wa...
research
09/17/2018

A convex formulation for high-dimensional sparse sliced inverse regression

Sliced inverse regression is a popular tool for sufficient dimension red...
research
12/20/2014

Outperforming Word2Vec on Analogy Tasks with Random Projections

We present a distributed vector representation based on a simplification...
research
05/12/2016

Subspace Perspective on Canonical Correlation Analysis: Dimension Reduction and Minimax Rates

Canonical correlation analysis (CCA) is a fundamental statistical tool f...
research
05/18/2021

High-Dimensional Sparse Single-Index Regression Via Hilbert-Schmidt Independence Criterion

Hilbert-Schmidt Independence Criterion (HSIC) has recently been used in ...
research
07/15/2022

Algorithmic Determination of the Combinatorial Structure of the Linear Regions of ReLU Neural Networks

We algorithmically determine the regions and facets of all dimensions of...

Please sign up or login with your details

Forgot password? Click here to reset