A Conditional Generative Chatbot using Transformer Model

06/03/2023
by   Nura Esfandiari, et al.
0

A Chatbot serves as a communication tool between a human user and a machine to achieve an appropriate answer based on the human input. In more recent approaches, a combination of Natural Language Processing and sequential models are used to build a generative Chatbot. The main challenge of these models is their sequential nature, which leads to less accurate results. To tackle this challenge, in this paper, a novel end-to-end architecture is proposed using conditional Wasserstein Generative Adversarial Networks and a transformer model for answer generation in Chatbots. While the generator of the proposed model consists of a full transformer model to generate an answer, the discriminator includes only the encoder part of a transformer model followed by a classifier. To the best of our knowledge, this is the first time that a generative Chatbot is proposed using the embedded transformer in both generator and discriminator models. Relying on the parallel computing of the transformer model, the results of the proposed model on the Cornell Movie-Dialog corpus and the Chit-Chat datasets confirm the superiority of the proposed model compared to state-of-the-art alternatives using different evaluation metrics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/15/2017

Improving Neural Machine Translation with Conditional Sequence Generative Adversarial Nets

This paper proposes an approach for applying GANs to NMT. We build a con...
research
02/21/2018

A Study into the similarity in generator and discriminator in GAN architecture

One popular generative model that has high-quality results is the Genera...
research
06/11/2018

Retinal Optic Disc Segmentation using Conditional Generative Adversarial Network

This paper proposed a retinal image segmentation method based on conditi...
research
05/22/2018

Robust Conditional Generative Adversarial Networks

Conditional generative adversarial networks (cGAN) have led to large imp...
research
09/14/2020

Contrastive Triple Extraction with Generative Transformer

Triple extraction is an essential task in information extraction for nat...
research
02/07/2022

HeadPosr: End-to-end Trainable Head Pose Estimation using Transformer Encoders

In this paper, HeadPosr is proposed to predict the head poses using a si...
research
05/03/2020

Transformer-based End-to-End Question Generation

Question Generation (QG) is an important task in Natural Language Proces...

Please sign up or login with your details

Forgot password? Click here to reset