SYNDICOM: Improving Conversational Commonsense with Error-Injection and Natural Language Feedback

09/18/2023
by   Christopher Richardson, et al.
0

Commonsense reasoning is a critical aspect of human communication. Despite recent advances in conversational AI driven by large language models, commonsense reasoning remains a challenging task. In this work, we introduce SYNDICOM - a method for improving commonsense in dialogue response generation. SYNDICOM consists of two components. The first component is a dataset composed of commonsense dialogues created from a knowledge graph and synthesized into natural language. This dataset includes both valid and invalid responses to dialogue contexts, along with natural language feedback (NLF) for the invalid responses. The second contribution is a two-step procedure: training a model to predict natural language feedback (NLF) for invalid responses, and then training a response generation model conditioned on the predicted NLF, the invalid response, and the dialogue. SYNDICOM is scalable and does not require reinforcement learning. Empirical results on three tasks are evaluated using a broad range of metrics. SYNDICOM achieves a relative improvement of 53 ChatGPT on ROUGE1, and human evaluators prefer SYNDICOM over ChatGPT 57 time. We will publicly release the code and the full dataset.

READ FULL TEXT
research
02/15/2023

Commonsense Reasoning for Conversational AI: A Survey of the State of the Art

Large, transformer-based pretrained language models like BERT, GPT, and ...
research
06/01/2021

CIDER: Commonsense Inference for Dialogue Explanation and Reasoning

Commonsense inference to understand and explain human language is a fund...
research
05/12/2023

ACCENT: An Automatic Event Commonsense Evaluation Metric for Open-Domain Dialogue Systems

Commonsense reasoning is omnipresent in human communications and thus is...
research
09/13/2019

Learning Household Task Knowledge from WikiHow Descriptions

Commonsense procedural knowledge is important for AI agents and robots t...
research
10/14/2020

Learning Improvised Chatbots from Adversarial Modifications of Natural Language Feedback

The ubiquitous nature of chatbots and their interaction with users gener...
research
05/22/2023

Can ChatGPT Defend the Truth? Automatic Dialectical Evaluation Elicits LLMs' Deficiencies in Reasoning

We explore testing the reasoning ability of large language models (LLMs)...
research
12/12/2021

ValueNet: A New Dataset for Human Value Driven Dialogue System

Building a socially intelligent agent involves many challenges, one of w...

Please sign up or login with your details

Forgot password? Click here to reset