Towards Building the Federated GPT: Federated Instruction Tuning

05/09/2023
by   Jianyi Zhang, et al.
0

While “instruction-tuned" generative large language models (LLMs) have demonstrated an impressive ability to generalize to new tasks, the training phases heavily rely on large amounts of diverse and high-quality instruction data (such as ChatGPT and GPT-4). Unfortunately, acquiring high-quality data, especially when it comes to human-written data, can pose significant challenges both in terms of cost and accessibility. Moreover, concerns related to privacy can further limit access to such data, making the process of obtaining it a complex and nuanced undertaking. Consequently, this hinders the generality of the tuned models and may restrict their effectiveness in certain contexts. To tackle this issue, our study introduces a new approach called Federated Instruction Tuning (FedIT), which leverages federated learning (FL) as the learning framework for the instruction tuning of LLMs. This marks the first exploration of FL-based instruction tuning for LLMs. This is especially important since text data is predominantly generated by end users. Therefore, it is imperative to design and adapt FL approaches to effectively leverage these users' diverse instructions stored on local devices, while preserving privacy and ensuring data security. In the current paper, by conducting widely used GPT-4 auto-evaluation, we demonstrate that by exploiting the heterogeneous and diverse sets of instructions on the client's end with the proposed framework FedIT, we improved the performance of LLMs compared to centralized training with only limited local instructions. Further, in this paper, we developed a Github repository named Shepherd. This repository offers a foundational framework for exploring federated fine-tuning of LLMs using heterogeneous instructions across diverse categories.

READ FULL TEXT
research
05/16/2023

Maybe Only 0.5 Training Data Instruction Tuning

Instruction tuning for large language models (LLMs) has gained attention...
research
08/12/2023

SLoRA: Federated Parameter Efficient Fine-Tuning of Language Models

Transfer learning via fine-tuning pre-trained transformer models has gai...
research
09/15/2023

FedJudge: Federated Legal Large Language Model

Large Language Models (LLMs) have gained prominence in the field of Lega...
research
07/14/2023

Federated Learning-Empowered AI-Generated Content in Wireless Networks

Artificial intelligence generated content (AIGC) has emerged as a promis...
research
05/19/2023

Federated Foundation Models: Privacy-Preserving and Collaborative Learning for Large Models

Foundation Models (FMs), such as BERT, GPT, ViT, and CLIP, have demonstr...
research
08/24/2023

Harnessing the Power of David against Goliath: Exploring Instruction Data Generation without Using Closed-Source Models

Instruction tuning is instrumental in enabling Large Language Models (LL...
research
07/20/2023

LLM Censorship: A Machine Learning Challenge or a Computer Security Problem?

Large language models (LLMs) have exhibited impressive capabilities in c...

Please sign up or login with your details

Forgot password? Click here to reset