Data-Efficient Cross-Lingual Transfer with Language-Specific Subnetworks

10/31/2022
by   Rochelle Choenni, et al.
0

Large multilingual language models typically share their parameters across all languages, which enables cross-lingual task transfer, but learning can also be hindered when training updates from different languages are in conflict. In this paper, we propose novel methods for using language-specific subnetworks, which control cross-lingual parameter sharing, to reduce conflicts and increase positive transfer during fine-tuning. We introduce dynamic subnetworks, which are jointly updated with the model, and we combine our methods with meta-learning, an established, but complementary, technique for improving cross-lingual transfer. Finally, we provide extensive analyses of how each of our methods affects the models.

READ FULL TEXT

page 7

page 8

research
04/08/2022

Marvelous Agglutinative Language Effect on Cross Lingual Transfer Learning

As for multilingual language models, it is important to select languages...
research
05/22/2023

How do languages influence each other? Studying cross-lingual data sharing during LLM fine-tuning

Multilingual large language models (MLLMs) are jointly trained on data f...
research
06/05/2023

Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers

Without any explicit cross-lingual training data, multilingual language ...
research
05/16/2021

The interplay between language similarity and script on a novel multi-layer Algerian dialect corpus

Recent years have seen a rise in interest for cross-lingual transfer bet...
research
05/23/2023

MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages

In this paper, we present MasakhaPOS, the largest part-of-speech (POS) d...
research
05/03/2023

Identifying the Correlation Between Language Distance and Cross-Lingual Transfer in a Multilingual Representation Space

Prior research has investigated the impact of various linguistic feature...
research
01/13/2023

Improving Generalization of Adapter-Based Cross-lingual Transfer with Scheduled Unfreezing

Standard fine-tuning of language models typically performs well on in-di...

Please sign up or login with your details

Forgot password? Click here to reset