A Comparative Study on Vocabulary Reduction for Phrase Table Smoothing

01/06/2019
by   Yunsu Kim, et al.
0

This work systematically analyzes the smoothing effect of vocabulary reduction for phrase translation models. We extensively compare various word-level vocabularies to show that the performance of smoothing is not significantly affected by the choice of vocabulary. This result provides empirical evidence that the standard phrase translation model is extremely sparse. Our experiments also reveal that vocabulary reduction is more effective for smoothing large-scale phrase tables.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2016

Vocabulary Manipulation for Neural Machine Translation

In order to capture rich language phenomena, neural machine translation ...
research
03/06/2022

Focus on the Target's Vocabulary: Masked Label Smoothing for Machine Translation

Label smoothing and vocabulary sharing are two widely used techniques in...
research
01/06/2019

Unsupervised Training for Large Vocabulary Translation Using Sparse Lexicon and Word Classes

We address for the first time unsupervised training for a translation ta...
research
12/01/2015

Augmenting Phrase Table by Employing Lexicons for Pivot-based SMT

Pivot language is employed as a way to solve the data sparseness problem...
research
11/27/2017

Query-Adaptive R-CNN for Open-Vocabulary Object Detection and Retrieval

We address the problem of open-vocabulary object retrieval and localizat...
research
01/14/2016

Smoothing parameter estimation framework for IBM word alignment models

IBM models are very important word alignment models in Machine Translati...
research
11/01/2016

Faster decoding for subword level Phrase-based SMT between related languages

A common and effective way to train translation systems between related ...

Please sign up or login with your details

Forgot password? Click here to reset