Efficient coordinate-descent for orthogonal matrices through Givens rotations

12/02/2013
by   Uri Shalit, et al.
0

Optimizing over the set of orthogonal matrices is a central component in problems like sparse-PCA or tensor decomposition. Unfortunately, such optimization is hard since simple operations on orthogonal matrices easily break orthogonality, and correcting orthogonality usually costs a large amount of computation. Here we propose a framework for optimizing orthogonal matrices, that is the parallel of coordinate-descent in Euclidean spaces. It is based on Givens-rotations, a fast-to-compute operation that affects a small number of entries in the learned matrix, and preserves orthogonality. We show two applications of this approach: an algorithm for tensor decomposition that is used in learning mixture models, and an algorithm for sparse-PCA. We study the parameter regime where a Givens rotation approach converges faster and achieves a superior model on a genome-wide brain-wide mRNA expression dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/18/2019

Fast approximation of orthogonal matrices and application to PCA

We study the problem of approximating orthogonal matrices so that their ...
research
04/04/2019

VeST: Very Sparse Tucker Factorization of Large-Scale Tensors

Given a large tensor, how can we decompose it to sparse core tensor and ...
research
05/30/2021

Parallelized Computation and Backpropagation Under Angle-Parametrized Orthogonal Matrices

We present a methodology for parallel acceleration of learning in the pr...
research
09/04/2022

Structure and approximation properties of Laplacian-like matrices

Many of today's problems require techniques that involve the solution of...
research
01/13/2022

When geometry meets optimization theory: partially orthogonal tensors

Due to the multi-linearity of tensors, most algorithms for tensor optimi...
research
02/15/2021

Fast and accurate optimization on the orthogonal manifold without retraction

We consider the problem of minimizing a function over the manifold of or...
research
12/07/2020

SGD_Tucker: A Novel Stochastic Optimization Strategy for Parallel Sparse Tucker Decomposition

Sparse Tucker Decomposition (STD) algorithms learn a core tensor and a g...

Please sign up or login with your details

Forgot password? Click here to reset