Personalized Decentralized Multi-Task Learning Over Dynamic Communication Graphs

12/21/2022
by   Matin Mortaheb, et al.
0

Decentralized and federated learning algorithms face data heterogeneity as one of the biggest challenges, especially when users want to learn a specific task. Even when personalized headers are used concatenated to a shared network (PF-MTL), aggregating all the networks with a decentralized algorithm can result in performance degradation as a result of heterogeneity in the data. Our algorithm uses exchanged gradients to calculate the correlations among tasks automatically, and dynamically adjusts the communication graph to connect mutually beneficial tasks and isolate those that may negatively impact each other. This algorithm improves the learning performance and leads to faster convergence compared to the case where all clients are connected to each other regardless of their correlations. We conduct experiments on a synthetic Gaussian dataset and a large-scale celebrity attributes (CelebA) dataset. The experiment with the synthetic data illustrates that our proposed method is capable of detecting tasks that are positively and negatively correlated. Moreover, the results of the experiments with CelebA demonstrate that the proposed method may produce significantly faster training results than fully-connected networks.

READ FULL TEXT
research
03/24/2022

FedGradNorm: Personalized Federated Gradient-Normalized Multi-Task Learning

Multi-task learning (MTL) is a novel framework to learn several tasks si...
research
01/24/2019

Communication-Efficient and Decentralized Multi-Task Boosting while Learning the Collaboration Graph

We study the decentralized machine learning scenario where many users co...
research
12/14/2022

Hierarchical Over-the-Air FedGradNorm

Multi-task learning (MTL) is a learning paradigm to learn multiple relat...
research
06/23/2023

Synthetic data shuffling accelerates the convergence of federated learning under data heterogeneity

In federated learning, data heterogeneity is a critical challenge. A str...
research
05/24/2023

Towards More Suitable Personalization in Federated Learning via Decentralized Partial Model Training

Personalized federated learning (PFL) aims to produce the greatest perso...
research
01/31/2022

BEER: Fast O(1/T) Rate for Decentralized Nonconvex Optimization with Communication Compression

Communication efficiency has been widely recognized as the bottleneck fo...
research
06/16/2023

Structured Cooperative Learning with Graphical Model Priors

We study how to train personalized models for different tasks on decentr...

Please sign up or login with your details

Forgot password? Click here to reset