Near optimal sample complexity for matrix and tensor normal models via geodesic convexity

10/14/2021
by   Cole Franks, et al.
7

The matrix normal model, the family of Gaussian matrix-variate distributions whose covariance matrix is the Kronecker product of two lower dimensional factors, is frequently used to model matrix-variate data. The tensor normal model generalizes this family to Kronecker products of three or more factors. We study the estimation of the Kronecker factors of the covariance matrix in the matrix and tensor models. We show nonasymptotic bounds for the error achieved by the maximum likelihood estimator (MLE) in several natural metrics. In contrast to existing bounds, our results do not rely on the factors being well-conditioned or sparse. For the matrix normal model, all our bounds are minimax optimal up to logarithmic factors, and for the tensor normal model our bound for the largest factor and overall covariance matrix are minimax optimal up to constant factors provided there are enough samples for any estimator to obtain constant Frobenius error. In the same regimes as our sample complexity bounds, we show that an iterative procedure to compute the MLE known as the flip-flop algorithm converges linearly with high probability. Our main tool is geodesic strong convexity in the geometry on positive-definite matrices induced by the Fisher information metric. This strong convexity is determined by the expansion of certain random quantum channels. We also provide numerical evidence that combining the flip-flop algorithm with a simple shrinkage estimator can improve performance in the undersampled regime.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2022

An Improvement on the Hotelling T^2 Test Using the Ledoit-Wolf Nonlinear Shrinkage Estimator

Hotelling's T^2 test is a classical approach for discriminating the mean...
research
06/13/2019

Sparse Approximate Factor Estimation for High-Dimensional Covariance Matrices

We propose a novel estimation approach for the covariance matrix based o...
research
07/24/2020

Principal Regression for High Dimensional Covariance Matrices

This manuscript presents an approach to perform generalized linear regre...
research
05/18/2018

Distributionally Robust Inverse Covariance Estimation: The Wasserstein Shrinkage Estimator

We introduce a distributionally robust maximum likelihood estimation mod...
research
11/26/2018

Finite Time Analysis of Vector Autoregressive Models under Linear Restrictions

This paper develops a unified finite-time theory for the OLS estimation ...
research
07/29/2009

Cooperative Training for Attribute-Distributed Data: Trade-off Between Data Transmission and Performance

This paper introduces a modeling framework for distributed regression wi...
research
01/22/2020

Estimating the reach of a manifold via its convexity defect function

The reach of a submanifold is a crucial regularity parameter for manifol...

Please sign up or login with your details

Forgot password? Click here to reset