Adversarially-Trained Normalized Noisy-Feature Auto-Encoder for Text Generation

11/10/2018
by   Xiang Zhang, et al.
0

This article proposes Adversarially-Trained Normalized Noisy-Feature Auto-Encoder (ATNNFAE) for byte-level text generation. An ATNNFAE consists of an auto-encoder where the internal code is normalized on the unit sphere and corrupted by additive noise. Simultaneously, a replica of the decoder (sharing the same parameters as the AE decoder) is used as the generator and fed with random latent vectors. An adversarial discriminator is trained to distinguish training samples reconstructed from the AE from samples produced through the random-input generator, making the entire generator-discriminator path differentiable for discrete data like text. The combined effect of noise injection in the code and shared weights between the decoder and the generator can prevent the mode collapsing phenomenon commonly observed in GANs. Since perplexity cannot be applied to non-sequential text generation, we propose a new evaluation method using the total variance distance between frequencies of hash-coded byte-level n-grams (NGTVD). NGTVD is a single benchmark that can characterize both the quality and the diversity of the generated texts. Experiments are offered in 6 large-scale datasets in Arabic, Chinese and English, with comparisons against n-gram baselines and recurrent neural networks (RNNs). Ablation study on both the noise level and the discriminator is performed. We find that RNNs have trouble competing with the n-gram baselines, and the ATNNFAE results are generally competitive.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/06/2018

Byte-Level Recursive Convolutional Auto-Encoder for Text

This article proposes to auto-encode text at byte-level using convolutio...
research
10/05/2018

Learning to Encode Text as Human-Readable Summaries using Generative Adversarial Networks

Auto-encoders compress input data into a latent-space representation and...
research
01/31/2020

Self-Adversarial Learning with Comparative Discrimination for Text Generation

Conventional Generative Adversarial Networks (GANs) for text generation ...
research
08/18/2023

MATLABER: Material-Aware Text-to-3D via LAtent BRDF auto-EncodeR

Based on powerful text-to-image diffusion models, text-to-3D generation ...
research
06/22/2020

Efficient text generation of user-defined topic using generative adversarial networks

This study focused on efficient text generation using generative adversa...
research
03/22/2021

SparseGAN: Sparse Generative Adversarial Network for Text Generation

It is still a challenging task to learn a neural text generation model u...
research
02/04/2021

Data-to-text Generation with Macro Planning

Recent approaches to data-to-text generation have adopted the very succe...

Please sign up or login with your details

Forgot password? Click here to reset