Measuring Catastrophic Forgetting in Cross-Lingual Transfer Paradigms: Exploring Tuning Strategies

09/12/2023
by   Boshko Koloski, et al.
0

The cross-lingual transfer is a promising technique to solve tasks in less-resourced languages. In this empirical study, we compare two fine-tuning approaches combined with zero-shot and full-shot learning approaches for large language models in a cross-lingual setting. As fine-tuning strategies, we compare parameter-efficient adapter methods with fine-tuning of all parameters. As cross-lingual transfer strategies, we compare the intermediate-training (IT) that uses each language sequentially and cross-lingual validation (CLV) that uses a target language already in the validation phase of fine-tuning. We assess the success of transfer and the extent of catastrophic forgetting in a source language due to cross-lingual transfer, i.e., how much previously acquired knowledge is lost when we learn new information in a different language. The results on two different classification problems, hate speech detection and product reviews, each containing datasets in several languages, show that the IT cross-lingual strategy outperforms CLV for the target language. Our findings indicate that, in the majority of cases, the CLV strategy demonstrates superior retention of knowledge in the base language (English) compared to the IT strategy, when evaluating catastrophic forgetting in multiple cross-lingual transfers.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/04/2023

DiTTO: A Feature Representation Imitation Approach for Improving Cross-Lingual Transfer

Zero-shot cross-lingual transfer is promising, however has been shown to...
research
01/13/2023

Improving Generalization of Adapter-Based Cross-lingual Transfer with Scheduled Unfreezing

Standard fine-tuning of language models typically performs well on in-di...
research
05/12/2023

Prompt Learning to Mitigate Catastrophic Forgetting in Cross-lingual Transfer for Open-domain Dialogue Generation

Dialogue systems for non-English languages have long been under-explored...
research
05/22/2023

How do languages influence each other? Studying cross-lingual data sharing during LLM fine-tuning

Multilingual large language models (MLLMs) are jointly trained on data f...
research
10/14/2021

Composable Sparse Fine-Tuning for Cross-Lingual Transfer

Fine-tuning all parameters of a pre-trained model has become the mainstr...
research
05/23/2023

Instruct-Align: Teaching Novel Languages with to LLMs through Alignment-based Cross-Lingual Instruction

Instruction-tuned large language models (LLMs) have shown remarkable gen...
research
06/05/2023

Cross-Lingual Transfer with Target Language-Ready Task Adapters

Adapters have emerged as a modular and parameter-efficient approach to (...

Please sign up or login with your details

Forgot password? Click here to reset