XeroAlign: Zero-Shot Cross-lingual Transformer Alignment

05/06/2021
by   Milan Gritta, et al.
0

The introduction of pretrained cross-lingual language models brought decisive improvements to multilingual NLP tasks. However, the lack of labelled task data necessitates a variety of methods aiming to close the gap to high-resource languages. Zero-shot methods in particular, often use translated task data as a training signal to bridge the performance gap between the source and target language(s). We introduce XeroAlign, a simple method for task-specific alignment of cross-lingual pretrained transformers such as XLM-R. XeroAlign uses translated task data to encourage the model to generate similar sentence embeddings for different languages. The XeroAligned XLM-R, called XLM-RA, shows strong improvements over the baseline models to achieve state-of-the-art zero-shot results on three multilingual natural language understanding tasks. XLM-RA's text classification accuracy exceeds that of XLM-R trained with labelled data and performs on par with state-of-the-art models on a cross-lingual adversarial paraphrasing task.

READ FULL TEXT
research
07/15/2023

Is Prompt-Based Finetuning Always Better than Vanilla Finetuning? Insights from Cross-Lingual Language Understanding

Multilingual pretrained language models (MPLMs) have demonstrated substa...
research
09/10/2020

FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding

Large-scale cross-lingual language models (LM), such as mBERT, Unicoder ...
research
04/19/2019

Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT

Pretrained contextual representation models (Peters et al., 2018; Devlin...
research
09/30/2020

Cross-lingual Spoken Language Understanding with Regularized Representation Alignment

Despite the promising results of current cross-lingual models for spoken...
research
09/09/2021

Subword Mapping and Anchoring across Languages

State-of-the-art multilingual systems rely on shared vocabularies that s...
research
10/11/2020

TransQuest at WMT2020: Sentence-Level Direct Assessment

This paper presents the team TransQuest's participation in Sentence-Leve...
research
04/04/2022

Aligned Weight Regularizers for Pruning Pretrained Neural Networks

While various avenues of research have been explored for iterative pruni...

Please sign up or login with your details

Forgot password? Click here to reset