How Do Multilingual Encoders Learn Cross-lingual Representation?

07/12/2022
by   Shijie Wu, et al.
0

NLP systems typically require support for more than one language. As different languages have different amounts of supervision, cross-lingual transfer benefits languages with little to no training data by transferring from other languages. From an engineering perspective, multilingual NLP benefits development and maintenance by serving multiple languages with a single system. Both cross-lingual transfer and multilingual NLP rely on cross-lingual representations serving as the foundation. As BERT revolutionized representation learning and NLP, it also revolutionized cross-lingual representations and cross-lingual transfer. Multilingual BERT was released as a replacement for single-language BERT, trained with Wikipedia data in 104 languages. Surprisingly, without any explicit cross-lingual signal, multilingual BERT learns cross-lingual representations in addition to representations for individual languages. This thesis first shows such surprising cross-lingual effectiveness compared against prior art on various tasks. Naturally, it raises a set of questions, most notably how do these multilingual encoders learn cross-lingual representations. In exploring these questions, this thesis will analyze the behavior of multilingual models in a variety of settings on high and low resource languages. We also look at how to inject different cross-lingual signals into multilingual encoders, and the optimization behavior of cross-lingual transfer with these models. Together, they provide a better understanding of multilingual encoders on cross-lingual transfer. Our findings will lead us to suggested improvements to multilingual encoders and cross-lingual transfer.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2020

A Study of Cross-Lingual Ability and Language-specific Information in Multilingual BERT

Recently, multilingual BERT works remarkably well on cross-lingual trans...
research
05/23/2023

Pixel Representations for Multilingual Translation and Data-efficient Cross-lingual Transfer

We introduce and demonstrate how to effectively train multilingual machi...
research
03/24/2020

XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization

Much recent progress in applications of machine learning models to NLP h...
research
05/02/2020

Gender Bias in Multilingual Embeddings and Cross-Lingual Transfer

Multilingual representations embed words from many languages into a sing...
research
10/24/2020

Cross-neutralising: Probing for joint encoding of linguistic information in multilingual models

Multilingual sentence encoders are widely used to transfer NLP models ac...
research
03/31/2020

Understanding Cross-Lingual Syntactic Transfer in Multilingual Recurrent Neural Networks

It is now established that modern neural language models can be successf...
research
08/05/2021

EENLP: Cross-lingual Eastern European NLP Index

This report presents the results of the EENLP project, done as a part of...

Please sign up or login with your details

Forgot password? Click here to reset