Homogenizing Non-IID datasets via In-Distribution Knowledge Distillation for Decentralized Learning

04/09/2023
by   Deepak Ravikumar, et al.
0

Decentralized learning enables serverless training of deep neural networks (DNNs) in a distributed manner on multiple nodes. This allows for the use of large datasets, as well as the ability to train with a wide variety of data sources. However, one of the key challenges with decentralized learning is heterogeneity in the data distribution across the nodes. In this paper, we propose In-Distribution Knowledge Distillation (IDKD) to address the challenge of heterogeneous data distribution. The goal of IDKD is to homogenize the data distribution across the nodes. While such data homogenization can be achieved by exchanging data among the nodes sacrificing privacy, IDKD achieves the same objective using a common public dataset across nodes without breaking the privacy constraint. This public dataset is different from the training dataset and is used to distill the knowledge from each node and communicate it to its neighbors through the generated labels. With traditional knowledge distillation, the generalization of the distilled model is reduced because all the public dataset samples are used irrespective of their similarity to the local dataset. Thus, we introduce an Out-of-Distribution (OoD) detector at each node to label a subset of the public dataset that maps close to the local training data distribution. Finally, only labels corresponding to these subsets are exchanged among the nodes and with appropriate label averaging each node is finetuned on these data subsets along with its local data. Our experiments on multiple image classification datasets and graph topologies show that the proposed IDKD scheme is more effective than traditional knowledge distillation and achieves state-of-the-art generalization performance on heterogeneously distributed data with minimal communication overhead.

READ FULL TEXT
research
05/02/2022

FedDKD: Federated Learning with Decentralized Knowledge Distillation

The performance of federated learning in neural networks is generally in...
research
03/21/2017

Knowledge distillation using unlabeled mismatched images

Current approaches for Knowledge Distillation (KD) either directly use t...
research
11/28/2022

Decentralized Learning with Multi-Headed Distillation

Decentralized learning with private data is a central problem in machine...
research
07/29/2023

The effect of network topologies on fully decentralized learning: a preliminary investigation

In a decentralized machine learning system, data is typically partitione...
research
11/11/2020

Real-Time Decentralized knowledge Transfer at the Edge

Proliferation of edge networks creates islands of learning agents workin...
research
04/03/2021

Knowledge Distillation For Wireless Edge Learning

In this paper, we propose a framework for predicting frame errors in the...
research
11/17/2022

Sub-Graph Learning for Spatiotemporal Forecasting via Knowledge Distillation

One of the challenges in studying the interactions in large graphs is to...

Please sign up or login with your details

Forgot password? Click here to reset