Continual-T0: Progressively Instructing 50+ Tasks to Language Models Without Forgetting

05/24/2022
by   Thomas Scialom, et al.
0

Recent work on large language models relies on the intuition that most natural language processing tasks can be described via natural language instructions. Language models trained on these instructions show strong zero-shot performance on several standard datasets. However, these models even though impressive still perform poorly on a wide range of tasks outside of their respective training and evaluation sets. To address this limitation, we argue that a model should be able to keep extending its knowledge and abilities, without forgetting previous skills. In spite of the limited success of Continual Learning we show that Language Models can be continual learners. We empirically investigate the reason for this success and conclude that Continual Learning emerges from self-supervision pre-training. Our resulting model Continual-T0 (CT0) is able to learn diverse new tasks, while still maintaining good performance on previous tasks, spanning remarkably through 70 datasets in total. Finally, we show that CT0 is able to combine instructions in ways it was never trained for, demonstrating some compositionality.

READ FULL TEXT
research
10/11/2022

Continual Training of Language Models for Few-Shot Learning

Recent work on applying large language models (LMs) achieves impressive ...
research
03/12/2023

Preventing Zero-Shot Transfer Degradation in Continual Learning of Vision-Language Models

Continual learning (CL) can help pre-trained vision-language models effi...
research
11/22/2019

Continual adaptation for efficient machine communication

To communicate with new partners in new contexts, humans rapidly form ne...
research
11/23/2022

Continual Learning of Natural Language Processing Tasks: A Survey

Continual learning (CL) is an emerging learning paradigm that aims to em...
research
03/09/2022

Memory Efficient Continual Learning for Neural Text Classification

Learning text classifiers based on pre-trained language models has becom...
research
07/11/2023

Towards Robust and Efficient Continual Language Learning

As the application space of language models continues to evolve, a natur...
research
03/02/2023

Semiparametric Language Models Are Scalable Continual Learners

Semiparametric language models (LMs) have shown promise in continuously ...

Please sign up or login with your details

Forgot password? Click here to reset