Cross-Lingual Relevance Transfer for Document Retrieval

11/08/2019
by   Peng Shi, et al.
0

Recent work has shown the surprising ability of multi-lingual BERT to serve as a zero-shot cross-lingual transfer model for a number of language processing tasks. We combine this finding with a similarly-recently proposal on sentence-level relevance modeling for document retrieval to demonstrate the ability of multi-lingual BERT to transfer models of relevance across languages. Experiments on test collections in five different languages from diverse language families (Chinese, Arabic, French, Hindi, and Bengali) show that models trained with English data improve ranking quality, without any special processing, both for (non-English) mono-lingual retrieval as well as cross-lingual retrieval.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/30/2018

Learning Cross-Lingual Sentence Representations via a Multi-task Dual-Encoder Model

Neural language models have been shown to achieve an impressive level of...
research
04/24/2020

Cross-lingual Information Retrieval with BERT

Multiple neural language models have been developed recently, e.g., BERT...
research
08/19/2021

Mr. TyDi: A Multi-lingual Benchmark for Dense Retrieval

We present Mr. TyDi, a multi-lingual benchmark dataset for mono-lingual ...
research
06/10/2021

Cross-lingual Emotion Detection

Emotion detection is of great importance for understanding humans. Const...
research
09/20/2021

BERT Cannot Align Characters

In previous work, it has been shown that BERT can adequately align cross...
research
12/21/2021

On Cross-Lingual Retrieval with Multilingual Text Encoders

In this work we present a systematic empirical study focused on the suit...
research
11/02/2020

Cross-Lingual Document Retrieval with Smooth Learning

Cross-lingual document search is an information retrieval task in which ...

Please sign up or login with your details

Forgot password? Click here to reset