Towards Language Agnostic Universal Representations

09/23/2018
by   Armen Aghajanyan, et al.
0

When a bilingual student learns to solve word problems in math, we expect the student to be able to solve these problem in both languages the student is fluent in,even if the math lessons were only taught in one language. However, current representations in machine learning are language dependent. In this work, we present a method to decouple the language from the problem by learning language agnostic representations and therefore allowing training a model in one language and applying to a different one in a zero shot fashion. We learn these representations by taking inspiration from linguistics and formalizing Universal Grammar as an optimization process (Chomsky, 2014; Montague, 1970). We demonstrate the capabilities of these representations by showing that the models trained on a single language using language agnostic representations achieve very similar accuracies in other languages.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/04/2021

On the ability of monolingual models to learn language-agnostic representations

Pretrained multilingual models have become a de facto default approach f...
research
05/26/2023

Gender Lost In Translation: How Bridging The Gap Between Languages Affects Gender Bias in Zero-Shot Multilingual Translation

Neural machine translation (NMT) models often suffer from gender biases ...
research
10/28/2022

Improving Zero-Shot Multilingual Translation with Universal Representations and Cross-Mappings

The many-to-many multilingual neural machine translation can translate b...
research
08/20/2020

Inducing Language-Agnostic Multilingual Representations

Multilingual representations have the potential to make cross-lingual sy...
research
11/15/2017

Tracking Typological Traits of Uralic Languages in Distributed Language Representations

Although linguistic typology has a long history, computational approache...
research
09/16/2021

Locating Language-Specific Information in Contextualized Embeddings

Multilingual pretrained language models (MPLMs) exhibit multilinguality ...
research
12/08/2014

Rediscovering the Alphabet - On the Innate Universal Grammar

Universal Grammar (UG) theory has been one of the most important researc...

Please sign up or login with your details

Forgot password? Click here to reset