To Beam Or Not To Beam: That is a Question of Cooperation for Language GANs

06/11/2021
by   Thomas Scialom, et al.
0

Due to the discrete nature of words, language GANs require to be optimized from rewards provided by discriminator networks, via reinforcement learning methods. This is a much harder setting than for continuous tasks, which enjoy gradient flows from discriminators to generators, usually leading to dramatic learning instabilities. However, we claim that this can be solved by making discriminator and generator networks cooperate to produce output sequences during training. These cooperative outputs, inherently built to obtain higher discrimination scores, not only provide denser rewards for training, but also form a more compact artificial set for discriminator training, hence improving its accuracy and stability. In this paper, we show that our SelfGAN framework, built on this cooperative principle, outperforms Teacher Forcing and obtains state-of-the-art results on two challenging tasks, Summarization and Question Generation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/28/2022

Generative Cooperative Networks for Natural Language Generation

Generative Adversarial Networks (GANs) have known a tremendous success f...
research
11/14/2022

Shared Loss between Generators of GANs

Generative adversarial networks are generative models that are capable o...
research
08/20/2019

ARAML: A Stable Adversarial Training Framework for Text Generation

Most of the existing generative adversarial networks (GAN) for text gene...
research
04/21/2022

6GAN: IPv6 Multi-Pattern Target Generation via Generative Adversarial Nets with Reinforcement Learning

Global IPv6 scanning has always been a challenge for researchers because...
research
07/24/2018

Improved Training with Curriculum GANs

In this paper we introduce Curriculum GANs, a curriculum learning strate...
research
05/22/2023

Cooperative Channel Capacity Learning

In this paper, the problem of determining the capacity of a communicatio...
research
10/01/2018

Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow

Adversarial learning methods have been proposed for a wide range of appl...

Please sign up or login with your details

Forgot password? Click here to reset