Neural Task Representations as Weak Supervision for Model Agnostic Cross-Lingual Transfer

11/02/2018
by   Sujay Kumar Jauhar, et al.
0

Natural language processing is heavily Anglo-centric, while the demand for models that work in languages other than English is greater than ever. Yet, the task of transferring a model from one language to another can be expensive in terms of annotation costs, engineering time and effort. In this paper, we present a general framework for easily and effectively transferring neural models from English to other languages. The framework, which relies on task representations as a form of weak supervision, is model and task agnostic, meaning that many existing neural architectures can be ported to other languages with minimal effort. The only requirement is unlabeled parallel data, and a loss defined over task representations. We evaluate our framework by transferring an English sentiment classifier to three different languages. On a battery of tests, we show that our models outperform a number of strong baselines and rival state-of-the-art results, which rely on more complex approaches and significantly more resources and data. Additionally, we find that the framework proposed in this paper is able to capture semantically rich and meaningful representations across languages, despite the lack of direct supervision.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/13/2018

XNLI: Evaluating Cross-lingual Sentence Representations

State-of-the-art natural language processing systems rely on supervision...
research
10/06/2020

Cross-Lingual Text Classification with Minimal Resources by Transferring a Sparse Teacher

Cross-lingual text classification alleviates the need for manually label...
research
05/30/2021

Learning Domain-Specialised Representations for Cross-Lingual Biomedical Entity Linking

Injecting external domain-specific knowledge (e.g., UMLS) into pretraine...
research
05/24/2018

A Corpus for Multilingual Document Classification in Eight Languages

Cross-lingual document classification aims at training a document classi...
research
04/09/2021

TransWiC at SemEval-2021 Task 2: Transformer-based Multilingual and Cross-lingual Word-in-Context Disambiguation

Identifying whether a word carries the same meaning or different meaning...
research
06/06/2016

Adversarial Deep Averaging Networks for Cross-Lingual Sentiment Classification

In recent years deep neural networks have achieved great success in sent...
research
01/17/2022

Handling Compounding in Mobile Keyboard Input

This paper proposes a framework to improve the typing experience of mobi...

Please sign up or login with your details

Forgot password? Click here to reset