mu-Forcing: Training Variational Recurrent Autoencoders for Text Generation

05/24/2019
by   Dayiheng Liu, et al.
0

It has been previously observed that training Variational Recurrent Autoencoders (VRAE) for text generation suffers from serious uninformative latent variables problem. The model would collapse into a plain language model that totally ignore the latent variables and can only generate repeating and dull samples. In this paper, we explore the reason behind this issue and propose an effective regularizer based approach to address it. The proposed method directly injects extra constraints on the posteriors of latent variables into the learning process of VRAE, which can flexibly and stably control the trade-off between the KL term and the reconstruction term, making the model learn dense and meaningful latent representations. The experimental results show that the proposed method outperforms several strong baselines and can make the model learn interpretable latent variables and generate diverse meaningful sentences. Furthermore, the proposed method can perform well without using other strategies, such as KL annealing.

READ FULL TEXT
research
07/13/2022

Fuse It More Deeply! A Variational Transformer with Layer-Wise Latent Variable Inference for Text Generation

The past several years have witnessed Variational Auto-Encoder's superio...
research
04/05/2022

latent-GLAT: Glancing at Latent Variables for Parallel Text Generation

Recently, parallel text generation has received widespread attention due...
research
07/28/2020

Novel Potential Inhibitors Against SARS-CoV-2 Using Artificial Intelligence

Abstract Since known approved drugs like liponavir and ritonavir failed ...
research
10/22/2022

Recurrence Boosts Diversity! Revisiting Recurrent Latent Variable in Transformer-Based Variational AutoEncoder for Diverse Text Generation

Variational Auto-Encoder (VAE) has been widely adopted in text generatio...
research
05/03/2020

Neural Data-to-Text Generation via Jointly Learning the Segmentation and Correspondence

The neural attention model has achieved great success in data-to-text ge...
research
04/10/2018

A Hierarchical Latent Structure for Variational Conversation Modeling

Variational autoencoders (VAE) combined with hierarchical RNNs have emer...
research
06/17/2020

Rethinking Semi-Supervised Learning in VAEs

We present an alternative approach to semi-supervision in variational au...

Please sign up or login with your details

Forgot password? Click here to reset