## 1 Introduction

Generative Adversarial Networks (GANs) [7] are used to train generative models in an adversarial setup, with a generator generating images that are trying to fool a discriminator whose role is to discriminate between real and synthetic images. GANs have shown good results and gained much attention when it comes to producing new images, however, language generation is an active and challenging area of research with GANs, mostly due to the non-differentiable nature of generating discrete symbols. Although GANs can accurately model complex distributions, they are known to be difficult to train due to instabilities caused by a difficult minimax optimization problem in a game theoretic situation, where the generator and discriminator aim to find a Nash Equilibrium.

Formally, the game between the generator G and the discriminator D is the minimax objective:

(1) |

where is the data distribution and is the model distribution implicitly defined by (the input

to the generator is sampled from some simple noise distribution p, such as the uniform distribution or a spherical Gaussian distribution). Wasserstein GANs (WGANs) with gradient penalty provide stable training of GANs and provide a better approach to reasonably train GANs.

[2]In this paper, we will be evaluating the use of Generative Adversarial Networks (GANs) for text generation and language modeling. In particular we look at two implementations

[3] and [5] using WGAN with gradient penalty. Furthermore, we provide some experimental results by training GANs with weaker regularization[8] and a modified gradient penalty that provides indications of better training stability of WGANs.## 2 Related work

Sai et. al [5]

have introduced a simple baseline that addresses the discrete output space problem without relying on gradient estimators and shows that it is able to achieve state-of-the-art results on a Chinese poem generation dataset and presented quantitative results on generating sentences from context-free and probabilistic context-free grammars, and qualitative language modeling results. A conditional version is also described that can generate sequences conditioned on sentence characteristics.

Ofir et. al [3]

have shown that recurrent neural networks can be trained to generate text with GANs from scratch using curriculum learning, by slowly teaching the model to generate sequences of increasing and variable length. They empirically show that their approach vastly improves the quality of generated sequences compared to a convolutional baseline.

Henning et. al [8] present theoretical arguments why using a weaker regularization term enforcing the Lipschitz constraint is preferable. These arguments are supported by experimental results on several data sets. For stable training of Wasserstein GANs, they propose to use the following penalty term to enforce the Lipschitz constraint that appears in the objective function:

(2) |

## 3 Data

We have used the One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling[1]. The project makes available a standard corpus of 0.8 billion words to train and evaluate language models. Duplicate sentences were removed, dropping the number of words from about 2.9 billion to about 0.8 billion. Words outside of the vocabulary were mapped to [unk] token, also part of the vocabulary. Sentence order was randomized, and the data was split into 100 disjoint partitions One such partition (1%) of the data was chosen as the held-out set. The benchmark is available as a code.google.com project at https://code.google.com/p/1-billion-word-language-modeling-benchmark/

; besides the scripts needed to rebuild the training/held-out data, it also makes available log-probability values for each word in each of ten held-out data sets, for each of the baseline n-gram models.

## 4 Methods

We have used the implementation in [3] and [5] as our baseline. In both cases, we have used the 1 billion word dataset. GAN based methods have often been critiqued for lacking a concrete evaluation strategy [9] and resort to some sort or subjective evaluation. For our evaluations, in [3] we generate 640 sequences from each model and measure %-IN-TEST-n, that is, the proportion of word n-grams from generated sequences that also appear in a held-out test set. We evaluate these metrics for n ∈ 1, 2, 3, 4., however for [5] we just rely on human understanding and thus is qualitative.

In both cases, we have used multiple hyper-parameter variations to find the most optimal set. For [3]

, we were able to improve the results by increasing the batch-size and increasing the number of epochs. Recently, Wasserstein GAN (WGAN) have been introduced which solve the vanishing gradient problem by clipping the weight values to lie in a cube. However, it has been shown that the gradient signals that generator receives might become unstable. An improved version uses weaker regularization for gradient penalty instead of clipping to force that double-sided gradient approaches. We have implemented this method and used it with a model trained based on

[3]. Training duration for GANs is unreasonably long, considering its reaches convergence at all. For [3] we have trained the network on four K80 GPUs for up to 14 days to run the variable length curriculum learning with teacher helping to generated sequences of length up to 25 with multiple hyper-parameters selections.## 5 Experiments/ Results/ Discussion

We used Amazon Elastic Compute Cloud and Microsoft Azure virtual machines to run our experiments. They provided the necessary CPU and GPU power to scale up our algorithms and run parallel processes. We evaluated our first model by generating sequences from the model and measuring ”percent in Test-N”, i.e. the proportion of word n-grams from generated sequences that also appear in a held-out test set. The goal is to measure the extent to which the generator is able to generate real words with local coherence. In the second model, we use human perception to decide on the quality of the produced text.

### 5.1 Language Generation with Recurrent Generative Adversarial Networks without Pre-training

This section follows the implementation on the lines of [3] where we train GRU based RNNs [4] to generate text using WGANs from scratch using curriculum learning, by teaching the model to generate sequences of increasing and variable length on the Billion word dataset [1]. We trained our model for up to 25 sequences on variable batch sizes, epochs, network layers, sequence lengths and observed the following results (Table 1)

Generated Samples | Unigrams | Bigrams | Trigrams | Quadgrams |
---|---|---|---|---|

”I have more than the man was” | ||||

”The NBC said so , Most of the” | 0.84 | 0.57 | 0.27 | 0.03 |

”Reid Maybe you got the more than” |

The proposed idea was to augment the WGAN loss by a regularization term that penalizes the deviation of the gradient norm of the critic with respect to its input from one (leading to a variant referred to as WGAN-GP, where GP stands for gradient penalty).[3]. However, we observed that increasing the batch size to 128 and using a weaker regularization term enforcing the Lipschitz constraint[8], led to highest accuracy across our experiments, and as early as sequence length 11 resulting to a Unigrams and Bigrams %-in-N accuracy of 85% and 57% respectively.

We also noticed an improvement in the stability and convergence of the GAN training when we used a modified gradient penalty [8] with the WGAN objective, as seen in (Figure 4) below. More specifically when we use equation to enforce the Lipschitz constraint that appears in the objective function instead of the previously considered approaches of clipping weights and of applying the stronger gradient penalty .

In addition to more stable learning behavior, the proposed regularization term leads to lower sensitivity to the value of the penalty weight (demonstrating smooth convergence and well-behaved critic scores throughout the whole training process for different values of )

### 5.2 Adversarial Generation of Natural Language

This section follows the implementation on the lines of [5]. We have access to a variety of generator and discriminator architectures including LSTM based RNN and 1D CNN. So far we got the best results by using LSTM for both generator and discriminator. We trained our model for different objective functions to generate 64 batches of 5-word sequences after every 50 iterations. As you see in Table 2, WGAN outperform GAN, and WGAN with GP outperforms WGAN.

We used both quantitative and qualitative measures to compare the results. As you see GAN makes frequent grammatical mistakes and repeats same words in a phrase. We see these errors a lot in the earlier stages of training, but even after many iterations, GAN keeps making those mistakes. WGAN has two advantages. It trains faster than GAN and produces more coherent results. Even though it still produce illegible phrases occasionally, it learns much faster not to repeat the same mistakes again. Adding gradient penalty makes WGAN more robust to errors and training instability. If not being set properly, the training process can be highly unstable and the loss starts bouncing around. We Used cross-validation on dev set to tune the hyper-parameters. Considering it takes a day to run each experiment, we used a small set of candidate values for hyper-parameters and used cross-validation on dev set to tune them.

The score indicates the fraction of all sentences produced during the training, which are not in the corpus. It is a rough estimate of how well the generator is performing.

After running the code for 5 epoch we got the following results:

Objective function | Samples | Score |
---|---|---|

”he only [unk] hard ?” | ||

GAN | ”about about in month ?” | 0.85 |

”the are are murder !” | ||

”like last doing monday arrested” | ||

WGAN | ”to all businesses won …” | 0.88 |

”they have were arrested ?” | ||

”your even example killed it” | ||

WGAN with GP | ”he could have started .” | 0.90 |

”finally cannot do parties ?” |

## 6 Conclusion/ Future Work

In conclusion, we show that GANs can produce original text with local coherency. Use of WGAN[10] is preferred over GAN as it trains faster, with higher stability, and produces more coherent phrases. Furthermore, we also show that using a weaker regularization term enforcing the Lipschitz constraint in the WGAN gradient penalty, has promised for more stable GAN training and smoother convergence.

In future work, we would like to explore more algorithms and techniques such as MaskGAN
[11], where we would like to extend the results of our experiments and evaluate its performance on MASKGAN which introduces an actor-critic conditional GAN that fills in missing text conditioned on the surrounding context. We would also like to train GANs on more coherent text corpus such as the Wikipedia dataset to generate language that is more globally coherent. As previously mentioned, GANs are challenging when it comes to training stability and convergence and the hyper-parameters are sensitive to small changes; With more time and computation power, we can run these algorithms for longer time and more variations.

## Acknowledgments

## References

- [1] Ciprian Chelba, Tomas Mikolov, Mike Schuster, Qi Ge, Thorsten Brants, Phillipp Koehn, Tony Robinson One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling. arXiv preprint arXiv:1312.3005v3, 2013.
- [2] Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron Courville Improved Training of Wasserstein GANs. arXiv preprint arXiv:1704.00028v3, 2017.
- [3] Ofir Press, Amir Bar, Ben Bogin, Jonathan Berant, Lior Wolf Language Generation with Recurrent Generative Adversarial Networks without Pre-training. arXiv preprint arXiv:1706.01399v3, 2017.
- [4] Kyunghyun Cho, Bart Van Merri¨enboer, Dzmitry Bahdanau and Yoshua Bengio On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259, 2014.
- [5] Sai Rajeswar, Sandeep Subramanian, Francis Dutil, Christopher Pal, Aaron Courville Adversarial Generation of Natural Language. arXiv preprint arXiv:1705.10929, 2017.
- [6] Samuel R. Bowman, Luke Vilnis, Oriol Vinyals, Andrew M. Dai, Rafal Jozefowicz and Samy Bengio Generating Sentences from a continuous space. arXiv preprint arXiv:1511.06349v2, 2015.
- [7] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville,Yoshua Bengio Generative Adversarial Networks. arXiv preprint arXiv:1406.2661v1, 2014.
- [8] Henning Petzka, Asja Fischer, Denis Lukovnicov On the regularization of Wasserstein GANs. arXiv preprint arXiv:1709.08894v2, 2017.
- [9] Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, Xi Chen Improved Techniques for Training GANs. arXiv preprint arXiv:1606.03498v1, 2016.
- [10] Martin Arjovsky, Soumith Chintala, Léon Bottou Wasserstein GAN. arXiv preprint arXiv:1701.07875v3, 2017.
- [11] William Fedus, Ian Goodfellow, Andrew M. Dai MaskGAN: Better Text Generation via Filling in the _____ . arXiv preprint arXiv:1801.07736v3, 2018.

Comments

There are no comments yet.