Communication-Efficient Federated Learning for Neural Machine Translation

12/12/2021
by   Tanya Roosta, et al.
0

Training neural machine translation (NMT) models in federated learning (FL) settings could be inefficient both computationally and communication-wise, due to the large size of translation engines as well as the multiple rounds of updates required to train clients and a central server. In this paper, we explore how to efficiently build NMT models in an FL setup by proposing a novel solution. In order to reduce the communication overhead, out of all neural layers we only exchange what we term "Controller" layers. Controllers are a small number of additional neural components connected to our pre-trained architectures. These new components are placed in between original layers. They act as liaisons to communicate with the central server and learn minimal information that is sufficient enough to update clients. We evaluated the performance of our models on five datasets from different domains to translate from German into English. We noted that the models equipped with Controllers preform on par with those trained in a central and non-FL setting. In addition, we observed a substantial reduction in the communication traffic of the FL pipeline, which is a direct consequence of using Controllers. Based on our experiments, Controller-based models are  6 times less expensive than their other peers. This reduction is significantly important when we consider the number of parameters in large models and it becomes even more critical when such parameters need to be exchanged for multiple rounds in FL settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2022

Training Mixed-Domain Translation Models via Federated Learning

Training mixed-domain translation models is a complex task that demands ...
research
02/23/2023

Federated Nearest Neighbor Machine Translation

To protect user privacy and meet legal regulations, federated learning (...
research
10/05/2021

Communication-Efficient Federated Learning with Binary Neural Networks

Federated learning (FL) is a privacy-preserving machine learning setting...
research
01/21/2021

Time-Correlated Sparsification for Communication-Efficient Federated Learning

Federated learning (FL) enables multiple clients to collaboratively trai...
research
05/21/2023

Communication Efficient Federated Learning for Multilingual Neural Machine Translation with Adapter

Federated Multilingual Neural Machine Translation (Fed-MNMT) has emerged...
research
02/02/2022

Federated Reinforcement Learning for Collective Navigation of Robotic Swarms

The recent advancement of Deep Reinforcement Learning (DRL) contributed ...
research
06/09/2023

Federated Learning You May Communicate Less Often!

We investigate the generalization error of statistical learning models i...

Please sign up or login with your details

Forgot password? Click here to reset