Low-complexity subspace-descent over symmetric positive definite manifold

05/03/2023
by   Yogesh Darmwal, et al.
0

This work puts forth low-complexity Riemannian subspace descent algorithms for the minimization of functions over the symmetric positive definite (SPD) manifold. Different from the existing Riemannian gradient descent variants, the proposed approach utilizes carefully chosen subspaces that allow the update to be written as a product of the Cholesky factor of the iterate and a sparse matrix. The resulting updates avoid the costly matrix operations like matrix exponentiation and dense matrix multiplication, which are generally required in almost all other Riemannian optimization algorithms on SPD manifold. We further identify a broad class of functions, arising in diverse applications, such as kernel matrix learning, covariance estimation of Gaussian distributions, maximum likelihood parameter estimation of elliptically contoured distributions, and parameter estimation in Gaussian mixture model problems, over which the Riemannian gradients can be calculated efficiently. The proposed uni-directional and multi-directional Riemannian subspace descent variants incur per-iteration complexities of 𝒪(n) and 𝒪(n^2) respectively, as compared to the 𝒪(n^3) or higher complexity incurred by all existing Riemannian gradient descent variants. The superior runtime and low per-iteration complexity of the proposed algorithms is also demonstrated via numerical tests on large-scale covariance estimation problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/18/2022

Riemannian Optimization for Variance Estimation in Linear Mixed Models

Variance parameter estimation in linear mixed models is a challenge for ...
research
02/01/2021

Probabilistic Learning Vector Quantization on Manifold of Symmetric Positive Definite Matrices

In this paper, we develop a new classification method for manifold-value...
research
03/12/2023

Gaussian kernels on non-simply-connected closed Riemannian manifolds are never positive definite

We show that the Gaussian kernel exp{-λ d_g^2(∙, ∙)} on any non-simply-c...
research
09/03/2019

Riemannian batch normalization for SPD neural networks

Covariance matrices have attracted attention for machine learning applic...
research
07/31/2017

Statistics on the (compact) Stiefel manifold: Theory and Applications

A Stiefel manifold of the compact type is often encountered in many fiel...
research
09/07/2022

Riemannian optimization for non-centered mixture of scaled Gaussian distributions

This paper studies the statistical model of the non-centered mixture of ...
research
12/09/2020

A Riemannian Block Coordinate Descent Method for Computing the Projection Robust Wasserstein Distance

The Wasserstein distance has become increasingly important in machine le...

Please sign up or login with your details

Forgot password? Click here to reset