Soft Layer Selection with Meta-Learning for Zero-Shot Cross-Lingual Transfer

07/21/2021
by   Weijia Xu, et al.
0

Multilingual pre-trained contextual embedding models (Devlin et al., 2019) have achieved impressive performance on zero-shot cross-lingual transfer tasks. Finding the most effective fine-tuning strategy to fine-tune these models on high-resource languages so that it transfers well to the zero-shot languages is a non-trivial task. In this paper, we propose a novel meta-optimizer to soft-select which layers of the pre-trained model to freeze during fine-tuning. We train the meta-optimizer by simulating the zero-shot transfer scenario. Results on cross-lingual natural language inference show that our approach improves over the simple fine-tuning baseline and X-MAML (Nooralahzadeh et al., 2020).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/22/2022

Prompt-Tuning Can Be Much Better Than Fine-Tuning on Cross-lingual Understanding With Multilingual Language Models

Pre-trained multilingual language models show significant performance ga...
research
06/16/2022

Zero-Shot AutoML with Pretrained Models

Given a new dataset D and a low compute budget, how should we choose a p...
research
04/26/2023

Neuro-symbolic Zero-Shot Code Cloning with Cross-Language Intermediate Representation

In this paper, we define a neuro-symbolic approach to address the task o...
research
05/22/2023

How do languages influence each other? Studying cross-lingual data sharing during LLM fine-tuning

Multilingual large language models (MLLMs) are jointly trained on data f...
research
10/14/2021

Composable Sparse Fine-Tuning for Cross-Lingual Transfer

Fine-tuning all parameters of a pre-trained model has become the mainstr...
research
03/30/2023

Fine-Tuning BERT with Character-Level Noise for Zero-Shot Transfer to Dialects and Closely-Related Languages

In this work, we induce character-level noise in various forms when fine...

Please sign up or login with your details

Forgot password? Click here to reset