Lexical Translation Model Using a Deep Neural Network Architecture

04/28/2015
by   Thanh-Le Ha, et al.
KIT
0

In this paper we combine the advantages of a model using global source sentence contexts, the Discriminative Word Lexicon, and neural networks. By using deep neural networks instead of the linear maximum entropy model in the Discriminative Word Lexicon models, we are able to leverage dependencies between different source words due to the non-linearity. Furthermore, the models for different target words can share parameters and therefore data sparsity problems are effectively reduced. By using this approach in a state-of-the-art translation system, we can improve the performance by up to 0.5 BLEU points for three different language pairs on the TED translation task.

READ FULL TEXT VIEW PDF

page 1

page 2

page 3

page 4

03/06/2015

Encoding Source Language with Convolutional Neural Network for Machine Translation

The recently proposed neural network joint model (NNJM) (Devlin et al., ...
08/17/2015

Effective Approaches to Attention-based Neural Machine Translation

An attentional mechanism has lately been used to improve neural machine ...
11/19/2019

A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

We propose a language-independent approach for improving statistical mac...
11/14/2015

Learning to Represent Words in Context with Multilingual Supervision

We present a neural network architecture based on bidirectional LSTMs to...
08/25/2018

Paraphrases as Foreign Languages in Multilingual Neural Machine Translation

Using paraphrases, the expression of the same semantic meaning in differ...
09/01/2018

Contextual Encoding for Translation Quality Estimation

The task of word-level quality estimation (QE) consists of taking a sour...
09/19/2019

DeepView: Visualizing the behavior of deep neural networks in a part of the data space

Machine learning models using deep architectures have been able to imple...

1 Introduction

Since the first attempt to statisical machine translation (SMT) [1], the approach has drawn much interest in the research community and huge improvements in translation quality have been achieved. Still, there are plenty of problems in SMT which should be addressed. One is that the translation decision depends on a quite small context.

In standard phrase-based statistical machine translation (PBMT) [2], the two main components are the translation and language models. The translation model is modeled by counting phrase pairs, which are sequences of words extracted from bilingual corpora. By using phrase segments instead of words, PBMT can exploit some local source and target contexts within those segments. But no context information outside the phrase pairs is used. In an -gram language model, only a context of up to target words is considered.

Several directions have been proposed to leverage information from wider contexts in the phrase-based SMT framework. For example, the Discriminative Word Lexicon (DWL) [3][4] exploits the occurence of all the words in the whole source sentence to predict the presence of words in the target sentence. This wider context information is encoded as features and employed in a discriminative framework. Hence, they train a maximum entropy (MaxEnt) model for each target word.

While this model can improve the translation quality in different conditions, MaxEnt models are linear classifiers. On the other hand, hierarchical non-linear classifiers can model dependencies between different source words better since they perform some abstraction over the input. Hence, introducing non-linearity into the modeling of the lexical translation could improve the quality. Moreover, since many pairs of source and target words co-occur only rarely, a way of sharing information between the different classifiers could improve the modeling as well.

In order to address these issues, we developed a discriminative lexical model based on deep neural networks. Since we train one neural network for all target words as a multivariate binary classifier, the model can share information between different target words. Furthermore, the probability is no longer a linear combination of weights depending on the surface source words. Thanks to the non-linearity, we are now able to exploit semantic dependencies among source words.

This paper is organized as follows. In Section 2, we review the previous works related to lexical translation methods as well as the translation modeling using neural networks. Then we describe our approach including the network architecture and its training procedures in Section 3. Section 4 provides experimental results of our translation systems for different language pairs using the proposed lexical translation model. Finally, the conclusions are drawn in Section  5.

2 Related work

Since the beginnings of SMT, several approaches to increase the context used for lexical decisions have been presented. When moving from word-based to phrase-based SMT [2][5], a big step in employing wider contexts into translation systems has been made. In PBMT, the lexical joint models allow us to use local source and target contexts in the form of phrases. Lately, advanced joint models have been proposed to either enhance the joint probability model between source and target sides or engage more suitable contexts.

The -gram based approach [6] directly models the joint probability of source and target sentences from the conditional probability of a current -gram pair givens sequences of previous bilingual -grams. In [7], this idea is introduced into the phrase-based MT approach. Thereby, parallel context over phrase boundaries can be used during the translation.

Standard phrase-based or -gram translation models are basically built upon statistical principles such as Maximum Entropy and smoothing techniques. Recently, joint models are learned using neural networks where non-linear translation relationships and semantic generalization of words can be performed [8]. Le et. al. [9] follow the -gram translation direction but model the conditional probability of a target word given the history of bilingual phrase pairs using a neural network architecture. They then use their model in a -best rescorer instead of in their -gram decoder. Devlin et. al. [10] add longer source contexts and renew the joint formula so that it can be included in a decoder rather than a -best rescoring module. Schwenk et. al. [11] calculate the conditional probability of a target phrase instead of a target word given a source phrase.

Although the aforementioned works essentially augment the joint translation model, they have an inherent limitation: only exploit local contexts. They estimate the joint model using sequences of words as the basic unit. On the other hand, there are several approaches utilizing global contexts. Motivated by Bangalore et. al

[12], Hasan et. al. [13] calculate the probability of a target word given two source words which do not necessarily belong to a phrase. Mauser et. al. [3] suggest another lexical translation approach, named Discriminative Word Lexicon (DWL), concentrating on predicting the presence of target words given the source words. Niehues et. al. [4] extend the model to employ the source and target contexts, but they used the same MaxEnt classifier for the task. Carpuat et. al. [14] is the most similar work to the DWL direction in terms of using the whole source sentence to perform the lexical choices of target words. They treat the selection process as a Word Sense Disambiguation (WSD) task, where target words or phrases are WSD senses. They extract a rich feature set from the source sentences, including source words, and input them into a WSD classifier. Still, the problem persists since they use the shallow classifiers for that task.

Considering the advantages of non-linear models mentioned before, we opt for using deep neural network architectures to learn the DWL. We take the advantages of the two directions. On one side, our model uses a non-linear classification method to leverage dependencies between different source sentences as well as its semantic generalization ability. On the other side, by employing the global contexts, our model can complement joint translation models which use the local contexts.

3 Discriminative lexical translation using deep neural networks

We will first review the original DWL approach described in [3] and [4]. Afterwards, we will describe the neural network architecture and training procedures proposed in this work. We will finish this section by describing the integration into the decoding process.

3.1 Original Discriminative Word Lexicon

In this approach, the DWL are modeled using a maximum entropy model to determine the probability of using a target word in the translation. Therefore, individual models for every target word are trained. Each model is trained to return the probability of this word given the input sentence.

The input of the model is the source sentence, thus, they need a way to represent the input sentence. This is done by representing the sentence as a bag of words and thereby ignoring the order of the words. In the MaxEnt model, they use an indicator feature for every input word. More formally, a given source sentence is represented by the features , with is the source vocabulary:

(1)

The models are trained on examples generated by the parallel training data. The labels for training the classifier of target word are defined as follows:

(2)

This model approximates the probability of a target word given the source sentence . We will discuss our alternative method using neural network to estimate those probabilities in the next section.

In [4], the source context is considered in a way that the sentence is no longer represented by a bag of words, but by a bag of ngrams. Using this representation, they could integrate the order information of the words, but the dimension of the input space is increased. We also adapt this extension to our model by encoding the bigrams and trigrams as ordinary words in the source vocabulary.

After inducing the probability for every word given the source sentence , these probabilities were combined into the probability of the whole target sentence given as described in Section 3.4.

3.2 General network architecture

After we reviewed the original DWL in the last section, we will now describe the neural network that replaces the MaxEnt model for calculating the probabilities .

Figure 1: FFNN architecture for learning lexical translation.

The input and output of our neural network-based DWL are the source and target sentences from which we would like to learn the lexical translation relationship. As in the original DWL approach, we represent each source sentence

as a binary column vector

with being the considered vocabulary of the source corpus. If a source word appears in that sentence , the value of the corresponding index in is , and otherwise. Hence, the source sentence representation should be a sparse vector, depending on the considered vocabulary . The same representation scheme is applied to the target sentence to get a sparse binary column vector with the considered target vocabulary .

As the Figure 1

depicts, our main neural network-based DWL architecture for learning lexical translation is a feed-forward neural network (FFNN) with three hidden layers. The matrix

connects the input layer to the first hidden layer. Two matrices and encodes the learned translation mapping between two compact global feature spaces of the source and target contexts. And the matrix computes the lexical translation output. , , and are the number of units in the first, second and third hidden layers, respectively. The lexical translation distribution of the words in the target sentence for a given source sentence is computed by a forward pass:

where:

and is the sigmoid function applied to the value in a column vector:

So the parameters of the network are:

To investigate the impact of the network configuration, we built a simpler architecture with only one hidden layer featuring the translation relationship between source and target sentences. We will refer this as the SimNNDWL in the comparison section later.

3.3 Network training

In neural network training, for each instance, which is comprised of a sentence pair , we maximize the similarity between the conditional probability to either 1 or 0 depending on the appearance of the corresponding word in the target sentence . The neural network operates as a multivariate classifier which gives the probabilistic score for a binary decision of independent variables, i.e the appearances of target words. Here we minimize the cross entropy error function between the binary target sentence vector and the output of the network :

We train the network by back-propagating the error based on the gradient descent principle. The error gradient for the weights between the last layer and the output is calculated as:

The error gradient for the weights between the other layers is calculated based on the error gradients for activation values from the previous layers:

Then the weight matrices are batch-updated after each epoch:

where:

  • is the number of training instances.

  • is the learning rate of the network.

  • is the weight matrix of the layer after epochs of training.

3.4 Sentence-level lexical translation scoring

With the independence assumption among target words, the target probabilities are combined to form the sentence-level lexical translation score:

(3)

where is the set of all target words appearing in the target sentence .

In Equation 3, we need to update the lexical translation score only if a new word appears in the hypothesis. That means we do not take into account the frequency of words but multiply the probability of one word only once even if the word occurs several times in the sentence. Other models in our translation system, however, will restrict overusing a particular word. Furthermore, to keep track of which words whose probabilities have been calculated already, additional book keeping would be required. In order to avoid those difficulties, we come up with the following approximation given is the length of the target sentence :

(4)

In order to speed up the calculation of the target word probabilities, we pre-calculate all probabilities for a given source sentence prior to translations. In a naive approach we would need to pre-calculate the probabilities for all possible target words given the source sentence. This would lead to a very slow calculations. Therefore, we first define the target vocabulary of a source sentence as the vocabulary comprised of the respective words from the phrase pairs matching to the source sentence. Using this definition, we only need to pre-calculate the probabilities of all words in the target side of the phrase table and not all target words in the whole corpus. And we can calculate the score for every phrase pair even before starting with the translation.

4 Experiments

In this section, we describe the translation system we use for the experiments, the configurations of the NNDWL and the results of those experiments.

4.1 System description

The system we use as our baseline is a state-of-the-art translation system for English to French without any DWL. To the baseline system, we add several DWL components trained on different corpora as independent features in the log-linear framework utilized by our in-house phrase-based decoder.

The system is trained on the EPPS, NC, Common Crawl, Giga corpora and TED talks[15]. The monolingual data we used to train language models includes the corresponding monolingual parts of those parallel corpora plus News Shuffle and Gigaword. The data is preprocessed and the phrase table is built using the scripts from the Moses package [16]. We adapt the general, big corpora to the in-domain TED data using the Backoff approach described in [17]. Adaptation is also conducted for the monolingual data. We train a 4-gram language model using the SRILM toolkit [18]. In addition, several non-word language models are included to capture the dependencies between source and target words and reduce the impact of data sparsity. We use a bilingual language model as described in [7] as well as a cluster language model based on word classes generated by the MKCLS algorithm [19]. Short-range reordering is performed as a preprocessing step as described in [20].

Our in-house phrase-based decoder is used to search for the best solutions among translation hypotheses and the optimization of the 13 to 17 features, depending on the settings we use, is performed using Minimum Error Rate Training [21]. The weights are optimized and tested on two separate sets of TED talks. The development set consists of 903 sentences containing 20k words. The test set consists of 1686 sentences containing 33k words.

We investigate the impact of our approach by employing different configurations of the neural networks described in details in the following section. We then evaluate those configurations not only for EnglishFrench but also for EnglishChinese and GermanEnglish with similar translation system setups.

Our NNDWL models are trained on a small subset of the mentioned training corpora, mainly the TED data. Although the TED corpus is quite small compared to the overall training data, it is very important since it matches best the test data. In order to speed up the process of testing different configurations, we therefore train the NNDWL only on this corpus except for the comparison reported in Section 4.3.4. The statistics of the training and validation data for the NNDWL are shown in Table 1.

En-Fr En-Zh De-En Training Sent. 149991 140006 130654 Tok. (avg.) 3.1m 3.3m 2.5m Validation Sent. 6153 8962 7430 Tok. (avg.) 125k 211k 142k

Table 1: Statistics of the corpora used to train NNDWL

4.2 Network configurations

In our main neural network architecture we proposed, the sizes of the hidden layers , , are , , , respectively. If we use the original source and target vocabularies, for the EnglishFrench direction trained on preprocessed TED 2013 data, includes words and includes words. Because of the non-linearity calculations through such a large network, the training is extremely time-consuming. In order to boost the efficiency, we limit the source and target vocabularies to the most frequent ones. All words outside the lists are treated as unknown words. We vary the size of the considered vocabularies from the values while keeping the sizes of the hidden layers the same (i.e. ). In preliminary experiments, this layout lead to the best performance. So we used this layout for the remaining of the paper.

The same calculation problem occurs with the source contexts, even more seriously due to the curse of dimentionality. Hence, we applied the same cut-off scheme to the source-side bigrams and trigrams with the most-frequent bigram and trigram numbers set at , and .

The simpler architecture SimNNDWL consisting of one 1000-unit hidden layer is compared to the main architecture with the same setup.

For training our proposed architecture, the gradient descent with a batch size of and a learning rate of is used. Gradients are calculated by averaging across a minibatch of training instances and the process is performed for 35 epochs. After each epoch, the current neural network model is evaluated on a separate validation set, and the model with the best performance on this set is utilized for calculating lexical translation scores afterwards. We regularize the models with the regularizer. As an alternative to the , we also experiment with the dropout technique [22]

, where the neurons in the last hidden layer are randomly dropped out with the probability of

. However, it did not help as indicated by its performance on the system later. The training is done on GPUs using the Theano Toolkit

[23].

4.3 Results

Here we report the results using different NNDWL configurations mainly for an EnglishFrench translation system. We also report the results using the best configurations for other language pairs.

4.3.1 Experiments with different vocabulary sizes

The results of the EnglishFrench translation system with NNDWL models trained with different vocabulary sizes are shown in Table 2.

System (En-Fr) BLEU BLEU Baseline 31.94 MaxEnt DWL 32.17 +0.23 NNDWL 500 32.06 +0.12 NNDWL 1000 32.37 +0.43 NNDWL 2000 32.38 +0.44 NNDWL 5000 32.07 +0.13 Full NNDWL 32.06 +0.12

Table 2: Results of the EnglishFrench NNDWL.

Varying the vocabulary sizes for both source and target sentences not only helps to dramatically reduce neural network training time but also affects the translation quality. In our experiments, neural networks with 1000- and 2000-most-frequent-word vocabularies show the biggest improvements with around 0.44 BLEU points in translating from English to French. They perform better than the DWL using the maximum entropy approach and the NNDWL with the whole source and target vocabularies.

While all NNDWL models achieve notable BLEU gains compared to the strong baseline, some of them are worse than the original MaxEnt model. It might be due to the fact that the original MaxEnt model uses the source contexts whereas the NNDWL models uses just the source words.

4.3.2 The impact of -gram source contexts

Tables 3 and 4 show the impact of bigrams and trigrams extracted from source sentences. We also vary the numbers of the bigrams and trigrams which appeared most often.

System (En-Fr) BLEU BLEU Baseline 31.94 NNDWL 2000 32.38 +0.44 NNDWL 2000 SC-200-100 32.35 +0.41 NNDWL 2000 SC-500-200 32.44 +0.50 NNDWL 2000 SC-1000-500 32.36 +0.42

Table 3: Results of the 2000-NNDWL with source contexts.

For the NNDWL model with 2000-most-frequent-word vocabularies, including source contexts helps in some cases and does not harm the translation performance in the other cases. With the 500 most-frequent bigrams and 200 most-frequent trigrams, we achieve the best improvements of BLEU points over the baseline.

System (En-Fr) BLEU BLEU Baseline 31.94 NNDWL 1000 32.37 +0.43 NNDWL 1000 SC-200-100 32.01 +0.07 NNDWL 1000 SC-500-200 32.23 +0.29 NNDWL 1000 SC-1000-500 32.39 +0.45

Table 4: Results of the 1000-NNDWL with source contexts.

The gains from adding source contexts to the 1000-vocabulary-size NNDWL model are not clearly observed as in the case of the 2000-vocabulary-size model. This might indicate that we should set the numbers of the source contexts to be proportional somehow with the size of the vocabularies.

4.3.3 The impact of using different architectures

System (En-Fr) BLEU BLEU
Baseline 31.94
NNDWL 1000 32.37 +0.43
SimNNDWL 1000 32.12 +0.18
NNDWL 2000 32.38 +0.44
SimNNDWL 2000 32.29 +0.35
NNDWL 5000 32.07 +0.13
SimNNDWL 5000 31.71 -0.23
Table 5: Results of NNDWL and SimNNDWL architectures.

Here we compare our main architecture with the simpler architecture SimNNDWL consisting of one 1000-unit hidden layer. While the SimNNDWL trains faster (157 hours vs. 202 hours for training EnglishFrench with the whole vocabularies), translation time performance is not significantly affected. Since there are decreases in BLEU score using SimNNDWL architecture as shown in Table 5, the deep architecture seems to have an advantage over the simple architecture. Hence, we stick with our main architecture for remaining experiments.

4.3.4 The impact of data used to train NNDWL models

We also train our NNDWL models on a bigger corpus concatinating EPPS, NC and TED. The results in Table 6 shows that using a bigger corpus does not improve the translation quality. The DWL models trained on in-domain data only, i.e. TED, perform similar or better than the models trained on more data but broader domains. This observation also holds true for original the MaxEnt DWL models reported in [24].

System (En-Fr) BLEU BLEU Baseline 31.94 NNDWL 1000 on TED 32.37 +0.43 NNDWL 1000 on EPPS+NC+TED 32.33 +0.39

Table 6: Results of the NNDWL trained on different corpora.

4.3.5 Other language pairs

We conducted the experiments with NNDWL models mainly on our English-to-French translation system in order to investigate the impact of our method on a strong baseline. However, we would like to inspect the effect of the DWL on language pairs with long-range dependencies or differences in word order.

For that purpose, we built similar NNDWL models and integrate them to our translation systems for other language pairs. Tables 7 and 8 show the results of EnglishChinese and GermanEnglish, respectively.

EnglishChinese

System (En-Zh) BLEU BLEU Baseline 17.18 MaxEnt DWL 16.78 -0.40 NNDWL 500 17.09 -0.09 NNDWL 1000 17.58 +0.40 NNDWL 1000 SC-200-100 17.63 +0.45 NNDWL 2000 17.26 +0.08 NNDWL 2000 SC-200-100 17.20 +0.02

Table 7: Results of the EnglishChinese NNDWL

In case of the EnglishChinese direction, the NNDWL significantly improves the translation quality, with an increment of 0.45 BLEU points over the baseline. That best BLEU gain comes from the NNDWL with 1000-most-frequent-word vocabularies and the source contexts containing 200 bigrams and 100 trigrams.

GermanEnglish

In case of the GermanEnglish direction, the NNDWL also helps to gain 0.34 BLEU points over the baseline with the best model (i.e. 2000 most-frequent-word vocabularies with source contexts). However, the improvements is not notably different compared to the original MaxEnt DWL.

System (De-En) BLEU BLEU Baseline 29.70 MaxEnt DWL 29.95 +0.25 NNDWL 500 29.82 +0.12 NNDWL 1000 29.92 +0.22 NNDWL 2000 29.95 +0.25 NNDWL 2000 SC-500-200 30.04 +0.34 NNDWL 5000 29.89 +0.19

Table 8: Results of the GermanEnglish NNDWL

5 Conclusion

In this paper we described a deep neural network approach for DWL modeling and the integration into a standard phrase-based translation system. Using neural networks as a non-linear classifier for DWL enables the ability of learning the abstract representation of global contexts and their dependencies. We investigated various network configurations on different language pairs. When we deployed our best NNDWL model as a feature in our decoder, it helps to improve up to 0.5 BLEU points compared to a very strong baseline.

Our NNDWL does not require linguistic resources nor feature engineering. Thus, it can easily be ported to new languages. Furthermore, the probability calculation can be done in a preprocessing step. Therefore, the new model would not significantly slow down the translation process. Although we do not feature linguistic resources in our NNDWL, they can be useful in modeling the translation probability of the languages from which they are avalaible. In future work we will try to integrate linguistic features into the model. Moreover, context vector of words might be helpful in further reducing the data sparseness problem.

6 Acknowledgments

The research leading to these results has received funding from the European Union Seventh Framework Programme (FP7/2007-2013) under grant agreement n 287658.

References

  • [1] P. F. Brown, V. J. D. Pietra, S. A. D. Pietra, and R. L. Mercer, “The Mathematics of Statistical Machine Translation: Parameter Estimation,” Computational Linguistics, vol. 19, no. 2, pp. 263–311, 1993.
  • [2] P. Koehn, F. J. Och, and D. Marcu, “Statistical Phrase-Based Translation,” in Proceedings of the 2003 Conference of the HLT/NAACL, 2003.
  • [3] A. Mauser, S. Hasan, and H. Ney, “Extending Statistical Machine Translation with Discriminative and Trigger-based Lexicon Models,” in

    Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 1 - Volume 1

    , ser. EMNLP ’09, Singapore, 2009.
  • [4] J. Niehues and A. Waibel, “An MT Error-driven Discriminative Word Lexicon using Sentence Structure Features,” in Proceedings of the Eighth Workshop on Statistical Machine Translation, 2013, pp. 512–520.
  • [5] F. J. Och and H. Ney, “The Alignment Template Approach to Statistical Machine Translation,” Computational Linguistics, vol. 30, no. 4, pp. 417–449, 2004.
  • [6]

    J. B. Marino, R. E. Banchs, J. M. Crego, A. de Gispert, P. Lambert, J. A. Fonollosa, and M. R. Costa-Jussà, “N-gram-based Machine Translation,”

    Computational Linguistics, vol. 32, no. 4, pp. 527–549, 2006.
  • [7] J. Niehues, T. Herrmann, S. Vogel, and A. Waibel, “Wider Context by Using Bilingual Language Models in Machine Translation,” in Proceedings of the Sixth Workshop on Statistical Machine Translation, Edinburgh, Scotland, 2011.
  • [8] T. Mikolov, W.-t. Yih, and G. Zweig, “Linguistic Regularities in Continuous Space Word Representations.” in Proceedings of HLT-NAACL, 2013, pp. 746–751.
  • [9] H.-S. Le, A. Allauzen, and F. Yvon, “Continuous Space Translation Models with Neural Networks,” in Proceedings of the 2012 Conference of the NAACL-HLT, Montréal, Canada, 2012.
  • [10] J. Devlin, R. Zbib, Z. Huang, T. Lamar, R. Schwartz, and J. Makhoul, “Fast and Robust Neural Network Joint Models for Statistical Machine Translation,” in Proceedings of the Association for Computational Linguistics (ACL), Baltimore, 2014.
  • [11] H. Schwenk, “Continuous Space Translation Models for Phrase-based Statistical Machine Translation.” in COLING (Posters), 2012, pp. 1071–1080.
  • [12] S. Bangalore, P. Haffner, and S. Kanthak, “Statistical Machine Translation through Global Lexical Selection and Sentence Reconstruction,” in Proceedings of 45 ACL), Prague, Czech Republic, 2007, pp. 152–159.
  • [13] S. Hasan, J. Ganitkevitch, H. Ney, and J. Andrés-Ferrer, “Triplet Lexicon Models for Statistical Machine Translation,” in Proceedings of the Conference on Empirical Methods in Natural Language Processing.   Association for Computational Linguistics, 2008, pp. 372–381.
  • [14] M. Carpuat and D. Wu, “Improving Statistical Machine Translation Using Word Sense Disambiguation.” in Proceedings of EMNLP-CoNLL, vol. 7, 2007, pp. 61–72.
  • [15] M. Cettolo, C. Girardi, and M. Federico, “WIT: Web Inventory of Transcribed and Translated Talks, year = 2012,” in Proceedings of the 16 Conference of the European Association for Machine Translation (EAMT), Trento, Italy, May, pp. 261–268.
  • [16] P. Koehn, H. Hoang, A. Birch, C. Callison-Burch, M. Federico, N. Bertoldi, B. Cowan, W. Shen, C. Moran, R. Zens, C. Dyer, O. Bojar, A. Constantin, and E. Herbst, “Moses: Open Source Toolkit for Statistical Machine Translation,” in ACL 2007, Demonstration Session, Prague, Czech Republic, June 2007.
  • [17] J. Niehues and A. Waibel, “Detailed Analysis of Different Strategies for Phrase Table Adaptation in SMT,” Proceedings of the Tenth Conference of the Association for Machine Translation in the America (AMTA), 2012.
  • [18] A. Stolcke, “SRILM – An Extensible Language Modeling Toolkit.” in Proc. of ICSLP, Denver, Colorado, USA, 2002.
  • [19] F. J. Och, “An Efficient Method for Determining Bilingual Word Classes.” in EACL’99, 1999.
  • [20] K. Rottmann and S. Vogel, “Word Reordering in Statistical Machine Translation with a POS-Based Distortion Model,” in TMI, Skövde, Sweden, 2007.
  • [21] F. J. Och, “Minimum Error Rate Training in Statistical Machine Translation,” in 41 Annual Meeting of the Association for Computational Linguistics (ACL), Sapporo, Japan, 2003.
  • [22] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov, “Improving Neural Networks by Preventing Co-adaptation of Feature Detectors,” arXiv preprint arXiv:1207.0580, 2012.
  • [23] J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu, G. Desjardins, J. Turian, D. Warde-Farley, and Y. Bengio, “Theano: a CPU and GPU Math Expression Compiler,” in Proceedings of the Python for Scientific Computing Conference (SciPy), June 2010.
  • [24] T.-L. Ha, T. Herrmann, J. Niehues, M. Mediani, E. Cho, Y. Zhang, I. Slawik, and A. Waibel, “The KIT Translation Systems for IWSLT 2013,” in Proceedings of the 2013 International Workshop on Spoken Language Translation (IWSLT), 2013.