Lattice minimum Bayes-risk (LMBR) decoding has been applied successfully to translation lattices in traditional SMT to improve translation performance of a single system [Kumar and Byrne2004, Tromble et al.2008, Blackwood et al.2010]. However, minimum Bayes-risk (MBR) decoding is also a very powerful framework for combining diverse systems [Sim et al.2007, de Gispert et al.2009]. Therefore, we study combining traditional SMT and NMT in a hybrid decoding scheme based on MBR. We argue that MBR-based methods in their present form are not well-suited for NMT because of the following reasons:
NMT decoding is computationally expensive. Therefore, it is difficult to collect the statistics needed for risk calculation for NMT.
The Bayes-risk in SMT is usually defined for complete translations. Therefore, the risk computation needs to be restructured in order to integrate it in an NMT decoder which builds up hypotheses from left to right.
To address these challenges, we use a special loss function which is computationally tractable as it avoids using NMT scores for risk calculation. We show how to reformulate the original LMBR decision rule for using it in a word-based NMT decoder which is not restricted to an
To address these challenges, we use a special loss function which is computationally tractable as it avoids using NMT scores for risk calculation. We show how to reformulate the original LMBR decision rule for using it in a word-based NMT decoder which is not restricted to an-best list or a lattice. Our hybrid system outperforms lattice rescoring on multiple data sets for English-German and Japanese-English. We report similar gains from applying our method to subword-unit-based NMT rather than word-based NMT.
2 Combining NMT and SMT by Minimising the Lattice Bayes-risk
We propose to collect statistics for MBR from a potentially large translation lattice generated with SMT, and use the -gram posteriors as additional score in NMT decoding. The LMBR decision rule used by Tromble et al. lmbr-tromble has the form
where is the hypothesis space of possible translations, is the evidence space for computing the Bayes-risk, and is the set of all -grams in (typically, ). In this work, our evidence space is a translation lattice generated with SMT. The function counts how often -gram occurs in translation . denotes the path posterior probability of
denotes the path posterior probability ofin . Our aim is to integrate these -gram posteriors into the NMT decoder since they correlate well with the presence of -grams in reference translations [de Gispert et al.2013]. We call the quantity to be maximised the evidence which corresponds to the (negative) Bayes-risk which is normally minimised in MBR decoding. We emphasize that this risk can be computed for any translation hypothesis and not only those produced by the SMT system.
NMT assigns a probability to a translation of source sentence via a left-to-right factorisation based on the chain rule:
via a left-to-right factorisation based on the chain rule:
where is a neural network using the hidden state of the decoder network and the context vector
is a neural network using the hidden state of the decoder network
and the context vectorwhich encodes relevant parts of the source sentence [Bahdanau et al.2015].111We refer to Bahdanau et al. bahdanau for a full discussion of attention-based NMT. can also represent an ensemble of NMT systems in which case the scores of the individual systems are multiplied together to form a single distribution. Applying the LMBR decision rule in Eq. 1 directly to NMT would involve computing for all translations in the evidence space. In case of LMBR this is equivalent to rescoring the entire translation lattice exhaustively with NMT. However, this is not feasible even for small lattices because the evaluation of is computationally very expensive. Therefore, we propose to calculate the Bayes-risk over SMT translation lattices using only pure SMT scores, and bias the NMT decoder towards low-risk hypotheses. Our final combined decision rule is
If contains a word not in the NMT vocabulary, the NMT model provides a score and updates its decoder state as for an unknown word. We note that can be computed even if is not in the SMT lattice. Therefore, Eq. 3 can be used to generate translations outside the SMT search space. We further note that Eq. 3 can be derived as an instance of LMBR under a modified loss function.
3 Left-to-right Decoding
Beam search is often used for NMT because the factorisation in Eq. 2 allows to build up hypotheses from left to right. In contrast, our definition of the evidence in Eq. 1 contains a sum over the (unordered) set of all -grams. However, we can rewrite our objective function in Eq. 3 in a way which makes it easy to use with beam search.
for -grams up to order 4. This form lends itself naturally to beam search: at each time step, we add to the previous partial hypothesis score both the log-likelihood of the last token according the NMT model, and the partial MBR gains from the current -gram history. Note that this is similar to applying (the exponentiated scores of) an interpolated language model based on
-gram history. Note that this is similar to applying (the exponentiated scores of) an interpolated language model based on-gram posteriors extracted from the SMT lattice. In the remainder of this paper, we will refer to decoding according Eq. 4 as MBR-based NMT.
|SMT baseline [de Gispert et al.2010, HiFST]||18.9||21.2||26.0|
|Single NMT (word)||Pure NMT||17.7||19.6||23.1|
|5-Ensemble NMT (word)||Pure NMT||19.4||21.8||25.4|
|Single NMT (BPE)||Pure NMT||19.6||21.9||24.6|
|3-Ensemble NMT (BPE)||Pure NMT||21.0||23.4||27.0|
4 Efficient -gram Posterior Calculation
The risk computation in our approach is based on posterior probabilities for -grams which we extract from the SMT translation lattice . is defined as the sum of the path probabilities of paths in containing [Blackwood et al.2010, Eq. 2]:
We use the framework of Blackwood et al. lmbr-blackwood based on -gram mapping and path counting transducers to efficiently pre-compute all non-zero values of . Complete enumeration of all -grams in a lattice is usually feasible even for very large lattices [Blackwood et al.2010]. Additionally, for all these -grams , we smooth by mixing it with the uniform distribution to flatten the distribution and increase the offset to
by mixing it with the uniform distribution to flatten the distribution and increase the offset to-grams which are not in the lattice.
5 Subword-unit-based NMT
Character-based or subword-unit-based NMT [Chitnis and DeNero2015, Sennrich et al.2016, Chung et al.2016, Luong and Manning2016, Costa-Jussà and Fonollosa2016, Ling et al.2015, Wu et al.2016] does not use isolated words as modelling units but applies a finer grained tokenization scheme. One of the main motivation for these approaches is to overcome the limited vocabulary in word-based NMT. We consider our hybrid system as an alternative way to fix NMT OOVs. However, our method can also be used with subword-unit-based NMT. In this work, we use byte pair encodings [Sennrich et al.2016, BPE] to test combining word-based SMT with subword-unit-based NMT via both lattice rescoring and MBR. First, we construct a finite state transducer (FST) which maps word sequences to BPE sequences. Then, we convert the word-based SMT lattices to BPE-based lattices by composing them with the mapping transducer and projecting the output tape using standard OpenFST operations [Allauzen et al.2007]. The converted lattices are used for extracting -gram posteriors as described in the previous sections. Note that even though the -grams are on the BPE level, their posteriors are computed from word-level SMT translation scores.
|SMT baseline [Neubig2013, Travatar]||19.5||22.2|
|Single NMT (word)||Pure NMT||20.3||22.5|
|6-Ensemble NMT (word)||Pure NMT||22.6||25.0|
|Single NMT (BPE)||Pure NMT||20.8||23.5|
|3-Ensemble NMT (BPE)||Pure NMT||23.3||25.9|
6 Experimental Setup
We test our approach on English-German (En-De) and Japanese-English (Ja-En). For En-De, we use the WMT news-test2014 (the filtered version) as a development set, and keep news-test2015 and news-test2016 as test sets. For Ja-En, we use the ASPEC corpus [Nakazawa et al.2016] to be strictly comparable to the evaluation done in the Workshop of Asian Translation (WAT).
The NMT systems are as described by Stahlberg et al. sgnmt using the Blocks and Theano frameworks
sgnmt using the Blocks and Theano frameworks[van Merriënboer et al.2015, Bastien et al.2012] with hyper-parameters as in [Bahdanau et al.2015] and a vocabulary size of 30k for Ja-En and 50k for En-De. We use the coverage penalty proposed by Wu et al. gnmt to improve the length and coverage of translations. Our final ensembles combine five (En-De) to six (Ja-En) independently trained NMT systems.
Our En-De SMT baseline is a hierarchical system based on the HiFST package444http://ucam-smt.github.io/ which produces rich output lattices. The system uses rules extracted as described by de Gispert et al. hifst-grammar and a 5-gram language model [Heafield et al.2013].
In Ja-En we use Travatar [Neubig2013], an open-source tree-to-string system. We provide the system with Japanese trees obtained using the Ckylark parser [Oda et al.2015] and train it on high-quality alignments as recommended by Neubig and Due smt-elements-tree2str. This system, which reproduces the results of the best submission in WAT 2014 [Neubig2014], is used to create a 10k-best list of hypotheses, which we convert into determinised and minimised FSAs for our work. Our Ja-En NMT models are trained on the same 500k training samples as the Travatar baseline.
The parameter is tuned by optimising the BLEU score on the validation set, and we set (). Using the BOBYQA algorithm [Powell2009] or lattice MERT [Macherey et al.2008] to optimise the -parameters independently did not yield improvements. The beam search implementation of the SGNMT decoder555http://ucam-smt.github.io/sgnmt/html/ [Stahlberg et al.2016b] is used in all our experiments. We set the beam size to 20 for En-De and 12 for Ja-En.
7 Results66footnotetext: Comparable to http://lotus.kuee.kyoto-u.ac.jp/WAT/evaluation/list.php?t=2
Our results are summarised in Tab. 1 and 2.777Instructions for reproducing our key results will be available upon publication at http://ucam-smt.github.io/sgnmt/html/tutorial.html Our approach outperforms both single NMT and SMT baselines by up to 3.4 BLEU for En-De and 2.8 BLEU for Ja-En. Ensembling yields further gains across all test sets both for the NMT baselines and our MBR-based hybrid systems. We see substantial gains from our MBR-based method over lattice rescoring for both single and ensembled NMT on all test sets and language pairs except En-De news-test2016. On Ja-En, we report 26.7 BLEU, second to only one system (as of February 2017) that uses a number of techniques such as minimum risk training and a much larger vocabulary size which could also be used in our framework.
Our word-level NMT baselines suffer from their limited vocabulary since we do not apply post-processing techniques like UNK-replace [Luong et al.2015]. Therefore, NMT with subword units (BPE) consistently outperforms them by a large margin. Lattice rescoring and MBR yield large gains for both BPE-based and word-based NMT. However, the performance difference between BPE- and word-level NMT diminishes with lattice rescoring and MBR decoding: rescoring with NMT often performs on the same level for both words and subword units, and MBR-based NMT is often even better with a word-level NMT baseline. This indicates that subword units are often not necessary when the hybrid system has access to a large word-level vocabulary like the SMT vocabulary.
Note that the BPE lattice rescoring system is constrained to produce words in the output vocabulary of the syntactic SMT system and is prevented from inventing new target language words out of combinations of subword units. MBR imposes a soft version of such a constraint by biasing the BPE-based system towards words in the SMT search space.
The hypotheses produced by our MBR-based method often differ from the translations in the baseline systems. For example, 77.8% of the translations from our best MBR-based system on Ja-En cannot be found in the SMT 10k-best list, and 78.0% do not match the translation from the pure NMT 6-ensemble.888Up to NMT OOVs. This suggests that our MBR decoder is able to produce entirely new hypotheses, and that our method has a profound effect on the translations which goes beyond rescoring the SMT search space or fixing UNKs in the NMT output.
Tab. 1 also shows that rescoring is sensitive to the size of the -best list or lattice: rescoring the entire lattice instead of a 100-best list often yields a gain of 1 full BLEU point. In order to test our MBR-based method on small lattices, we compiled -best lists of varying sizes to lattices and extracted -gram posteriors from the reduced lattices. Fig. 1 shows that the -best list size has an impact on both methods. Rescoring a 10-best list already yields a large improvement of 1.2 BLEU. However, the hypotheses are still close to the SMT baseline. The MBR-based approach can make better use of small -best lists as it does not suffer this restriction. MBR-based combination on a 10-best list performs on about the same level as rescoring a 10,000-best list which demonstrates a practical advantage of MBR over rescoring.
8 Related Work
Combining the advantages of NMT and traditional SMT has received some attention in current research. A recent line of research attempts to integrate SMT-style translation tables into the NMT system [Zhang and Zong2016, Arthur et al.2016, He et al.2016]. Wang et al. nmt+smt-train interpolated NMT posteriors with word recommendations from SMT and jointly trained NMT together with a gating function which assigns the weight between SMT and NMT scores dynamically. Neubig et al. neubigAtwat2015 rescored -best lists from a syntax-based SMT system with NMT. Stahlberg et al. sgnmt restricted the NMT search space to a Hiero lattice and reported improvements over -best list rescoring. Stahlberg et al. editdistance combined Hiero and NMT via a loose coupling scheme based on composition of finite state transducers and translation lattices which takes the edit distance between translations into account. Our approach is similar to the latter one since it allows to divert from SMT and generate translations without derivations in the SMT system. This ability is crucial for NMT ensembles because SMT lattices are often too narrow for the NMT decoder [Stahlberg et al.2016a]. However, the method proposed by Stahlberg et al. editdistance insists on a monotone alignment between SMT and NMT translations to calculate the edit distance. This can be computationally expensive and not appropriate for MT where word reorderings are common. The MBR decoding described here does not have this shortcoming.
This paper discussed a novel method for blending NMT with traditional SMT by biasing NMT scores towards translations with low Bayes-risk with respect to the SMT lattice. We reported significant improvements of the new method over lattice rescoring on Japanese-English and English-German and showed that it can make good use even of very small lattices and -best lists.
In this work, we calculated the Bayes-risk over non-neural SMT lattices. In the future, we are planning to introduce neural models to the risk estimation while keeping the computational complexity under control, e.g. by using neural -gram language models [Bengio et al.2003, Vaswani et al.2013] or approximations of NMT scores [Lecorvé and Motlicek2012, Liu et al.2016] for -gram posterior calculation.
This work was supported by the U.K. Engineering and Physical Sciences Research Council (EPSRC grant EP/L027623/1).
We thank Graham Neubig for providing pre-trained parsing and alignment models, as well as scripts, to allow perfect reproduction of the NAIST WAT 2014 submission.
- [Allauzen et al.2007] Cyril Allauzen, Michael Riley, Johan Schalkwyk, Wojciech Skut, and Mehryar Mohri. 2007. Openfst: A general and efficient weighted finite-state transducer library. In International Conference on Implementation and Application of Automata, pages 11--23. Springer.
[Arthur et al.2016]
Philip Arthur, Graham Neubig, and Satoshi Nakamura.
Incorporating discrete translation lexicons into neural machine translation.In EMNLP, pages 1557--1567.
- [Bahdanau et al.2015] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In ICLR.
- [Bastien et al.2012] Frédéric Bastien, Pascal Lamblin, Razvan Pascanu, James Bergstra, Ian Goodfellow, Arnaud Bergeron, Nicolas Bouchard, David Warde-Farley, and Yoshua Bengio. 2012. Theano: new features and speed improvements. In NIPS.
[Bengio et al.2003]
Yoshua Bengio, Réjean Ducharme, Pascal Vincent, and Christian Jauvin.
A neural probabilistic language model.
Journal of machine learning research, 3:1137--1155.
- [Blackwood et al.2010] Graeme Blackwood, Adrià de Gispert, and William Byrne. 2010. Efficient path counting transducers for minimum Bayes-risk decoding of statistical machine translation lattices. In ACL, pages 27--32.
- [Chitnis and DeNero2015] Rohan Chitnis and John DeNero. 2015. Variable-length word encodings for neural translation models. In EMNLP, pages 2088--2093.
- [Chung et al.2016] Junyoung Chung, Kyunghyun Cho, and Yoshua Bengio. 2016. A character-level decoder without explicit segmentation for neural machine translation. In ACL.
- [Costa-Jussà and Fonollosa2016] Marta R. Costa-Jussà and José AR. Fonollosa. 2016. Character-based neural machine translation. In ACL.
- [de Gispert et al.2009] Adrià de Gispert, Sami Virpioja, Mikko Kurimo, and William Byrne. 2009. Minimum Bayes risk combination of translation hypotheses from alternative morphological decompositions. In HLT-NAACL, pages 73--76.
- [de Gispert et al.2010] Adrià de Gispert, Gonzalo Iglesias, Graeme Blackwood, Eduardo R Banga, and William Byrne. 2010. Hierarchical phrase-based translation with weighted finite-state transducers and shallow-n grammars. Computational Linguistics, 36(3):505--533.
- [de Gispert et al.2013] Adrià de Gispert, Graeme Blackwood, Gonzalo Iglesias, and William Byrne. 2013. N-gram posterior probability confidence measures for statistical machine translation: an empirical study. Machine Translation, 27(2):85--114.
- [He et al.2016] Wei He, Zhongjun He, Hua Wu, and Haifeng Wang. 2016. Improved neural machine translation with SMT features. In AAAI.
- [Heafield et al.2013] Kenneth Heafield, Ivan Pouzyrevsky, Jonathan H. Clark, and Philipp Koehn. 2013. Scalable modified Kneser-Ney language model estimation. In ACL, pages 690--696.
- [Kumar and Byrne2004] Shankar Kumar and William Byrne. 2004. Minimum Bayes-risk decoding for statistical machine translation. In HLT-NAACL.
[Lecorvé and Motlicek2012]
Gwénolé Lecorvé and Petr Motlicek.
Conversion of recurrent neural network language models to weighted finite state transducers for automatic speech recognition.Technical report, Idiap.
- [Li and Jurafsky2016] Jiwei Li and Dan Jurafsky. 2016. Mutual information and diverse decoding improve neural machine translation. arXiv preprint arXiv:1601.00372.
- [Ling et al.2015] Wang Ling, Isabel Trancoso, Chris Dyer, and Alan W. Black. 2015. Character-based neural machine translation. arXiv preprint arXiv:1511.04586.
- [Liu et al.2016] Xunying Liu, Xie Chen, Yongqiang Wang, Mark JF. Gales, and Philip C. Woodland. 2016. Two efficient lattice rescoring methods using recurrent neural network language models. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 24(8):1438--1449.
- [Luong and Manning2016] Minh-Thang Luong and Christopher D. Manning. 2016. Achieving open vocabulary neural machine translation with hybrid word-character models. In ACL.
- [Luong et al.2015] Minh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, and Wojciech Zaremba. 2015. Addressing the rare word problem in neural machine translation. In ACL.
- [Macherey et al.2008] Wolfgang Macherey, Franz Josef Och, Ignacio Thayer, and Jakob Uszkoreit. 2008. Lattice-based minimum error rate training for statistical machine translation. In EMNLP, pages 725--734.
- [Nakazawa et al.2016] Toshiaki Nakazawa, Manabu Yaguchi, Kiyotaka Uchimoto, Masao Utiyama, Eiichiro Sumita, Sadao Kurohashi, and Hitoshi Isahara. 2016. ASPEC: Asian scientific paper excerpt corpus. In LREC, pages 2204--2208.
- [Neubig and Duh2014] Graham Neubig and Kevin Duh. 2014. On the elements of an accurate tree-to-string machine translation system. In ACL, pages 143--149.
- [Neubig et al.2015] Graham Neubig, Makoto Morishita, and Satoshi Nakamura. 2015. Neural reranking improves subjective quality of machine translation: NAIST at WAT2015. In WAT, Kyoto, Japan.
- [Neubig2013] Graham Neubig. 2013. Travatar: A forest-to-string machine translation engine based on tree transducers. In ACL.
- [Neubig2014] Graham Neubig. 2014. Forest-to-string SMT for Asian language translation: NAIST at WAT 2014. In WAT.
- [Oda et al.2015] Yusuke Oda, Graham Neubig, Sakriani Sakti, Tomoki Toda, and Satoshi Nakamura. 2015. Ckylark: A more robust PCFG-LA parser. In NAACL, pages 41--45.
- [Powell2009] Michael JD. Powell. 2009. The BOBYQA algorithm for bound constrained optimization without derivatives. Cambridge NA Report NA2009/06, University of Cambridge, Cambridge.
- [Sennrich et al.2016] Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016. Neural machine translation of rare words with subword units. In ACL.
- [Sim et al.2007] Khe Chai Sim, William J. Byrne, Mark JF. Gales, Hichem Sahbi, and Philip C. Woodland. 2007. Consensus network decoding for statistical machine translation system combination. In ICASSP. IEEE.
- [Stahlberg et al.2016a] Felix Stahlberg, Eva Hasler, and Bill Byrne. 2016a. The edit distance transducer in action: The University of Cambridge English-German system at WMT16. In WMT.
- [Stahlberg et al.2016b] Felix Stahlberg, Eva Hasler, Aurelien Waite, and Bill Byrne. 2016b. Syntactically guided neural machine translation. In ACL.
- [Tromble et al.2008] Roy W. Tromble, Shankar Kumar, Franz Och, and Wolfgang Macherey. 2008. Lattice minimum Bayes-risk decoding for statistical machine translation. In EMNLP, pages 620--629.
[van Merriënboer et al.2015]
Bart van Merriënboer, Dzmitry Bahdanau, Vincent Dumoulin, Dmitriy
Serdyuk, David Warde-Farley, Jan Chorowski, and Yoshua Bengio.
Blocks and fuel: Frameworks for deep learning.CoRR.
- [Vaswani et al.2013] Ashish Vaswani, Yinggong Zhao, Victoria Fossum, and David Chiang. 2013. Decoding with large-scale neural language models improves translation. In EMNLP, pages 1387--1392.
- [Vijayakumar et al.2016] Ashwin K. Vijayakumar, Michael Cogswell, Ramprasath R. Selvaraju, Qing Sun, Stefan Lee, David Crandall, and Dhruv Batra. 2016. Diverse beam search: Decoding diverse solutions from neural sequence models. arXiv preprint arXiv:1610.02424.
- [Wang et al.2016] Xing Wang, Zhengdong Lu, Zhaopeng Tu, Hang Li, Deyi Xiong, and Min Zhang. 2016. Neural machine translation advised by statistical machine translation. CoRR, abs/1610.05150.
- [Wu et al.2016] Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. 2016. Google’s neural machine translation system: Bridging the gap between human and machine translation. CoRR, abs/1609.08144.
- [Zhang and Zong2016] Jiajun Zhang and Chengqing Zong. 2016. Bridging neural machine translation and bilingual dictionaries. arXiv preprint arXiv:1610.07272.