Federated NLP in Few-shot Scenarios

12/12/2022
by   Dongqi Cai, et al.
0

Natural language processing (NLP) sees rich mobile applications. To support various language understanding tasks, a foundation NLP model is often fine-tuned in a federated, privacy-preserving setting (FL). This process currently relies on at least hundreds of thousands of labeled training samples from mobile clients; yet mobile users often lack willingness or knowledge to label their data. Such an inadequacy of data labels is known as a few-shot scenario; it becomes the key blocker for mobile NLP applications. For the first time, this work investigates federated NLP in the few-shot scenario (FedFSL). By retrofitting algorithmic advances of pseudo labeling and prompt learning, we first establish a training pipeline that delivers competitive accuracy when only 0.05 labeled and the remaining is unlabeled. To instantiate the workflow, we further present a system FFNLP, addressing the high execution cost with novel designs. (1) Curriculum pacing, which injects pseudo labels to the training workflow at a rate commensurate to the learning progress; (2) Representational diversity, a mechanism for selecting the most learnable data, only for which pseudo labels will be generated; (3) Co-planning of a model's training depth and layer capacity. Together, these designs reduce the training delay, client energy, and network traffic by up to 46.0×, 41.2× and 3000.0×, respectively. Through algorithm/system co-design, FFNLP demonstrates that FL can apply to challenging settings where most training samples are unlabeled.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/26/2023

Low-Parameter Federated Learning with Large Language Models

We study few-shot Natural Language Understanding (NLU) tasks with Large ...
research
05/27/2022

Federated Semi-Supervised Learning with Prototypical Networks

With the increasing computing power of edge devices, Federated Learning ...
research
08/21/2021

SemiFed: Semi-supervised Federated Learning with Consistency and Pseudo-Labeling

Federated learning enables multiple clients, such as mobile phones and o...
research
12/16/2022

FewFedWeight: Few-shot Federated Learning Framework across Multiple NLP Tasks

Massively multi-task learning with large language models has recently ma...
research
12/12/2022

Collaborating Heterogeneous Natural Language Processing Tasks via Federated Learning

The increasing privacy concerns on personal private text data promote th...
research
06/14/2023

Fed-ZERO: Efficient Zero-shot Personalization with Federated Mixture of Experts

One of the goals in Federated Learning (FL) is to create personalized mo...
research
07/11/2022

Efficient NLP Inference at the Edge via Elastic Pipelining

Natural Language Processing (NLP) inference is seeing increasing adoptio...

Please sign up or login with your details

Forgot password? Click here to reset