Communication-Efficient Distributed Linear and Deep Generalized Canonical Correlation Analysis

09/25/2021
by   Sagar Shrestha, et al.
0

Classic and deep learning-based generalized canonical correlation analysis (GCCA) algorithms seek low-dimensional common representations of data entities from multiple “views” (e.g., audio and image) using linear transformations and neural networks, respectively. When the views are acquired and stored at different locations, organizations and edge devices, computing GCCA in a distributed, parallel and efficient manner is well-motivated. However, existing distributed GCCA algorithms may incur prohitively high communication overhead. This work puts forth a communication-efficient distributed framework for both linear and deep GCCA under the maximum variance (MAX-VAR) paradigm. The overhead issue is addressed by aggressively compressing (via quantization) the exchanging information between the distributed computing agents and a central controller. Compared to the unquantized version, the proposed algorithm consistently reduces the communication overhead by about 90% with virtually no loss in accuracy and convergence speed. Rigorous convergence analyses are also presented – which is a nontrivial effort since no existing generic result from quantized distributed optimization covers the special problem structure of GCCA. Our result shows that the proposed algorithms for both linear and deep GCCA converge to critical points in a sublinear rate, even under heavy quantization and stochastic approximations. In addition, it is shown that in the linear MAX-VAR case, the quantized algorithm approaches a global optimum in a geometric rate – if the computing agents' updates meet a certain accuracy level. Synthetic and real data experiments are used to showcase the effectiveness of the proposed approach.

READ FULL TEXT
research
05/31/2016

Scalable and Flexible Multiview MAX-VAR Canonical Correlation Analysis

Generalized canonical correlation analysis (GCCA) aims at finding latent...
research
04/02/2019

Nested Dithered Quantization for Communication Reduction in Distributed Training

In distributed training, the communication cost due to the transmission ...
research
03/25/2020

Generalized Canonical Correlation Analysis: A Subspace Intersection Approach

Generalized Canonical Correlation Analysis (GCCA) is an important tool t...
research
10/09/2019

High-Dimensional Stochastic Gradient Quantization for Communication-Efficient Edge Learning

Edge machine learning involves the deployment of learning algorithms at ...
research
04/24/2018

Structured SUMCOR Multiview Canonical Correlation Analysis for Large-Scale Data

The sum-of-correlations (SUMCOR) formulation of generalized canonical co...
research
04/23/2020

Sparse Generalized Canonical Correlation Analysis: Distributed Alternating Iteration based Approach

Sparse canonical correlation analysis (CCA) is a useful statistical tool...
research
06/02/2021

Hyperdimensional Computing for Efficient Distributed Classification with Randomized Neural Networks

In the supervised learning domain, considering the recent prevalence of ...

Please sign up or login with your details

Forgot password? Click here to reset