Distributionally Robust Reduced Rank Regression and Principal Component Analysis in High Dimensions

10/18/2018
by   Kean Ming Tan, et al.
0

We propose robust sparse reduced rank regression and robust sparse principal component analysis for analyzing large and complex high-dimensional data with heavy-tailed random noise. The proposed methods are based on convex relaxations of rank-and sparsity-constrained non-convex optimization problems, which are solved using the alternating direction method of multipliers (ADMM) algorithm. For robust sparse reduced rank regression, we establish non-asymptotic estimation error bounds under both Frobenius and nuclear norms, while existing results focus mostly on rank-selection and prediction consistency. Our theoretical results quantify the tradeoff between heavy-tailedness of the random noise and statistical bias. For random noise with bounded (1+δ)th moment with δ∈ (0,1), the rate of convergence is a function of δ, and is slower than the sub-Gaussian-type deviation bounds; for random noise with bounded second moment, we recover the results obtained under sub-Gaussian noise. Furthermore, the transition between the two regimes is smooth. For robust sparse principal component analysis, we propose to truncate the observed data, and show that this truncation will lead to consistent estimation of the eigenvectors. We then establish theoretical results similar to those of robust sparse reduced rank regression. We illustrate the performance of these methods via extensive numerical studies and two real data applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/01/2021

Sparse principal component analysis for high-dimensional stationary time series

We consider the sparse principal component analysis for high-dimensional...
research
12/25/2022

Mining the Factor Zoo: Estimation of Latent Factor Models with Sufficient Proxies

Latent factor model estimation typically relies on either using domain k...
research
01/17/2023

Sparse and Integrative Principal Component Analysis for Multiview Data

We consider dimension reduction of multiview data, which are emerging in...
research
08/20/2016

Visual Processing by a Unified Schatten-p Norm and ℓ_q Norm Regularized Principal Component Pursuit

In this paper, we propose a non-convex formulation to recover the authen...
research
12/15/2021

On Generalization and Computation of Tukey's Depth: Part II

This paper studies how to generalize Tukey's depth to problems defined i...
research
07/02/2019

An Iteratively Re-weighted Method for Problems with Sparsity-Inducing Norms

This work aims at solving the problems with intractable sparsity-inducin...
research
05/10/2023

Computationally Efficient and Statistically Optimal Robust High-Dimensional Linear Regression

High-dimensional linear regression under heavy-tailed noise or outlier c...

Please sign up or login with your details

Forgot password? Click here to reset