Why Does Zero-Shot Cross-Lingual Generation Fail? An Explanation and a Solution

05/27/2023
by   Tianjian Li, et al.
0

Zero-shot cross-lingual transfer is when a multilingual model is trained to perform a task in one language and then is applied to another language. Although the zero-shot cross-lingual transfer approach has achieved success in various classification tasks, its performance on natural language generation tasks falls short in quality and sometimes outputs an incorrect language. In our study, we show that the fine-tuning process learns language invariant representations, which is beneficial for classification tasks but harmful for generation tasks. Motivated by this, we propose a simple method to regularize the model from learning language invariant representations and a method to select model checkpoints without a development set in the target language, both resulting in better generation quality. Experiments on three semantically diverse generation tasks show that our method reduces the accidental translation problem by 68

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2022

Realistic Zero-Shot Cross-Lingual Transfer in Legal Topic Classification

We consider zero-shot cross-lingual transfer in legal topic classificati...
research
06/23/2023

Voicebox: Text-Guided Multilingual Universal Speech Generation at Scale

Large-scale generative models such as GPT and DALL-E have revolutionized...
research
06/11/2020

CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot Cross-Lingual NLP

Multi-lingual contextualized embeddings, such as multilingual-BERT (mBER...
research
10/10/2020

Zero-Shot Translation Quality Estimation with Explicit Cross-Lingual Patterns

This paper describes our submission of the WMT 2020 Shared Task on Sente...
research
05/26/2023

Free Lunch: Robust Cross-Lingual Transfer via Model Checkpoint Averaging

Massively multilingual language models have displayed strong performance...
research
01/24/2023

Cross-lingual German Biomedical Information Extraction: from Zero-shot to Human-in-the-Loop

This paper presents our project proposal for extracting biomedical infor...
research
10/23/2020

Unsupervised Cross-lingual Adaptation for Sequence Tagging and Beyond

Cross-lingual adaptation with multilingual pre-trained language models (...

Please sign up or login with your details

Forgot password? Click here to reset