Towards Making the Most of ChatGPT for Machine Translation

03/24/2023
by   Keqin Peng, et al.
0

ChatGPT shows remarkable capabilities for machine translation (MT). Several prior studies have shown that it achieves comparable results to commercial systems for high-resource languages, but lags behind in complex tasks, e.g, low-resource and distant-language-pairs translation. However, they usually adopt simple prompts which can not fully elicit the capability of ChatGPT. In this report, we aim to further mine ChatGPT's translation ability by revisiting several aspects: temperature, task information, and domain information, and correspondingly propose two (simple but effective) prompts: Task-Specific Prompts (TSP) and Domain-Specific Prompts (DSP). We show that: 1) The performance of ChatGPT depends largely on temperature, and a lower temperature usually can achieve better performance; 2) Emphasizing the task information further improves ChatGPT's performance, particularly in complex MT tasks; 3) Introducing domain information can elicit ChatGPT's generalization ability and improve its performance in the specific domain; 4) ChatGPT tends to generate hallucinations for non-English-centric MT tasks, which can be partially addressed by our proposed prompts but still need to be highlighted for the MT/NLP community. We also explore the effects of advanced in-context learning strategies and find a (negative but interesting) observation: the powerful chain-of-thought prompt leads to word-by-word translation behavior, thus bringing significant translation degradation.

READ FULL TEXT
research
03/31/2021

Domain-specific MT for Low-resource Languages: The case of Bambara-French

Translating to and from low-resource languages is a challenge for machin...
research
09/14/2023

ChatGPT MT: Competitive for High- (but not Low-) Resource Languages

Large language models (LLMs) implicitly learn to perform a range of lang...
research
07/03/2021

Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN

Despite their practical success, modern seq2seq architectures are unable...
research
06/11/2019

Translating Translationese: A Two-Step Approach to Unsupervised Machine Translation

Given a rough, word-by-word gloss of a source language sentence, target ...
research
05/17/2021

Ensemble-based Transfer Learning for Low-resource Machine Translation Quality Estimation

Quality Estimation (QE) of Machine Translation (MT) is a task to estimat...
research
09/10/2021

Dynamic Terminology Integration for COVID-19 and other Emerging Domains

The majority of language domains require prudent use of terminology to e...
research
09/20/2022

Vega-MT: The JD Explore Academy Translation System for WMT22

We describe the JD Explore Academy's submission of the WMT 2022 shared g...

Please sign up or login with your details

Forgot password? Click here to reset