Character-based NMT with Transformer

11/12/2019
by   Laurent Besacier, et al.
0

Character-based translation has several appealing advantages, but its performance is in general worse than a carefully tuned BPE baseline. In this paper we study the impact of character-based input and output with the Transformer architecture. In particular, our experiments on EN-DE show that character-based Transformer models are more robust than their BPE counterpart, both when translating noisy text, and when translating text from a different domain. To obtain comparable BLEU scores in clean, in-domain data and close the gap with BPE-based models we use known techniques to train deeper Transformer models.

READ FULL TEXT
research
04/29/2020

Towards Character-Level Transformer NMT by Finetuning Subword Systems

Applying the Transformer architecture on the character level usually req...
research
05/23/2022

Use of Transformer-Based Models for Word-Level Transliteration of the Book of the Dean of Lismore

The Book of the Dean of Lismore (BDL) is a 16th-century Scottish Gaelic ...
research
05/05/2023

Adapting Transformer Language Models for Predictive Typing in Brain-Computer Interfaces

Brain-computer interfaces (BCI) are an important mode of alternative and...
research
05/22/2020

Character-level Transformer-based Neural Machine Translation

Neural machine translation (NMT) is nowadays commonly applied at the sub...
research
03/10/2022

TrueType Transformer: Character and Font Style Recognition in Outline Format

We propose TrueType Transformer (T3), which can perform character and fo...
research
09/30/2021

SCIMAT: Science and Mathematics Dataset

In this work, we announce a comprehensive well curated and opensource da...
research
05/27/2022

Patching Leaks in the Charformer for Efficient Character-Level Generation

Character-based representations have important advantages over subword-b...

Please sign up or login with your details

Forgot password? Click here to reset