Understanding Crosslingual Transfer Mechanisms in Probabilistic Topic Modeling

10/13/2018
by   Shudong Hao, et al.
0

Probabilistic topic modeling is a popular choice as the first step of crosslingual tasks to enable knowledge transfer and extract multilingual features. While many multilingual topic models have been developed, their assumptions on the training corpus are quite varied, and it is not clear how well the models can be applied under various training conditions. In this paper, we systematically study the knowledge transfer mechanisms behind different multilingual topic models, and through a broad set of experiments with four models on ten languages, we provide empirical insights that can inform the selection and future development of multilingual topic models.

READ FULL TEXT
research
06/11/2018

Learning Multilingual Topics from Incomparable Corpus

Multilingual topic models enable crosslingual tasks by extracting consis...
research
05/09/2012

Multilingual Topic Models for Unaligned Text

We develop the multilingual topic model for unaligned text (MuTo), a pro...
research
04/26/2018

Lessons from the Bible on Modern Topics: Low-Resource Multilingual Topic Model Evaluation

Multilingual topic models enable document analysis across languages thro...
research
12/18/2017

Multilingual Topic Models

Scientific publications have evolved several features for mitigating voc...
research
10/24/2020

Cross-neutralising: Probing for joint encoding of linguistic information in multilingual models

Multilingual sentence encoders are widely used to transfer NLP models ac...
research
06/11/2023

Language Versatilists vs. Specialists: An Empirical Revisiting on Multilingual Transfer Ability

Multilingual transfer ability, which reflects how well the models fine-t...
research
04/28/2023

Training and Evaluation of a Multilingual Tokenizer for GPT-SW3

This paper provides a detailed discussion of the multilingual tokenizer ...

Please sign up or login with your details

Forgot password? Click here to reset