DeepAI
Log In Sign Up

Towards User-Driven Neural Machine Translation

06/11/2021
by   Huan Lin, et al.
0

A good translation should not only translate the original content semantically, but also incarnate personal traits of the original text. For a real-world neural machine translation (NMT) system, these user traits (e.g., topic preference, stylistic characteristics and expression habits) can be preserved in user behavior (e.g., historical inputs). However, current NMT systems marginally consider the user behavior due to: 1) the difficulty of modeling user portraits in zero-shot scenarios, and 2) the lack of user-behavior annotated parallel dataset. To fill this gap, we introduce a novel framework called user-driven NMT. Specifically, a cache-based module and a user-driven contrastive learning method are proposed to offer NMT the ability to capture potential user traits from their historical inputs under a zero-shot learning fashion. Furthermore, we contribute the first Chinese-English parallel corpus annotated with user behavior called UDT-Corpus. Experimental results confirm that the proposed user-driven NMT can generate user-specific translations.

READ FULL TEXT

page 1

page 2

page 3

page 4

11/21/2017

Effective Strategies in Zero-Shot Neural Machine Translation

In this paper, we proposed two strategies which can be applied to a mult...
05/25/2018

Zero-Shot Dual Machine Translation

Neural Machine Translation (NMT) systems rely on large amounts of parall...
01/27/2022

Learning How to Translate North Korean through South Korean

South and North Korea both use the Korean language. However, Korean NLP ...
11/02/2021

Zero-Shot Translation using Diffusion Models

In this work, we show a novel method for neural machine translation (NMT...
10/02/2021

Improving Zero-shot Multilingual Neural Machine Translation for Low-Resource Languages

Although the multilingual Neural Machine Translation(NMT), which extends...
01/10/2018

Translating Pro-Drop Languages with Reconstruction Models

Pronouns are frequently omitted in pro-drop languages, such as Chinese, ...