Cross-lingual Intermediate Fine-tuning improves Dialogue State Tracking

09/28/2021
by   Nikita Moghe, et al.
6

Recent progress in task-oriented neural dialogue systems is largely focused on a handful of languages, as annotation of training data is tedious and expensive. Machine translation has been used to make systems multilingual, but this can introduce a pipeline of errors. Another promising solution is using cross-lingual transfer learning through pretrained multilingual models. Existing methods train multilingual models with additional code-mixed task data or refine the cross-lingual representations through parallel ontologies. In this work, we enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models, where the multilingual models are fine-tuned with different but related data and/or tasks. Specifically, we use parallel and conversational movie subtitles datasets to design cross-lingual intermediate tasks suitable for downstream dialogue tasks. We use only 200K lines of parallel data for intermediate fine-tuning which is already available for 1782 language pairs. We test our approach on the cross-lingual dialogue state tracking task for the parallel MultiWoZ (English -> Chinese, Chinese -> English) and Multilingual WoZ (English -> German, English -> Italian) datasets. We achieve impressive improvements (> 20 parallel MultiWoZ dataset and the Multilingual WoZ dataset over the vanilla baseline with only 10 respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/05/2022

Parameter-Efficient Neural Reranking for Cross-Lingual and Multilingual Retrieval

State-of-the-art neural (re)rankers are notoriously data hungry which - ...
research
06/05/2023

Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers

Without any explicit cross-lingual training data, multilingual language ...
research
10/06/2019

Multilingual Dialogue Generation with Shared-Private Memory

Existing dialog systems are all monolingual, where features shared among...
research
04/03/2023

Efficiently Aligned Cross-Lingual Transfer Learning for Conversational Tasks using Prompt-Tuning

Cross-lingual transfer of language models trained on high-resource langu...
research
04/11/2022

Zero-shot Cross-lingual Conversational Semantic Role Labeling

While conversational semantic role labeling (CSRL) has shown its usefuln...
research
08/19/2018

XL-NBT: A Cross-lingual Neural Belief Tracking Framework

Task-oriented dialog systems are becoming pervasive, and many companies ...
research
10/13/2022

A Multi-dimensional Evaluation of Tokenizer-free Multilingual Pretrained Models

Recent work on tokenizer-free multilingual pretrained models show promis...

Please sign up or login with your details

Forgot password? Click here to reset