How "open" are the conversations with open-domain chatbots? A proposal for Speech Event based evaluation

11/24/2022
by   A. Seza Doğruöz, et al.
0

Open-domain chatbots are supposed to converse freely with humans without being restricted to a topic, task or domain. However, the boundaries and/or contents of open-domain conversations are not clear. To clarify the boundaries of "openness", we conduct two studies: First, we classify the types of "speech events" encountered in a chatbot evaluation data set (i.e., Meena by Google) and find that these conversations mainly cover the "small talk" category and exclude the other speech event categories encountered in real life human-human communication. Second, we conduct a small-scale pilot study to generate online conversations covering a wider range of speech event categories between two humans vs. a human and a state-of-the-art chatbot (i.e., Blender by Facebook). A human evaluation of these generated conversations indicates a preference for human-human conversations, since the human-chatbot conversations lack coherence in most speech event categories. Based on these results, we suggest (a) using the term "small talk" instead of "open-domain" for the current chatbots which are not that "open" in terms of conversational abilities yet, and (b) revising the evaluation methods to test the chatbot conversations against other speech events.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/17/2023

LED: A Dataset for Life Event Extraction from Dialogs

Lifelogging has gained more attention due to its wide applications, such...
research
04/12/2021

Building a Swedish Open-Domain Conversational Language Model

We present on-going work of evaluating the, to our knowledge, first larg...
research
01/31/2019

The Second Conversational Intelligence Challenge (ConvAI2)

We describe the setting and results of the ConvAI2 NeurIPS competition t...
research
08/19/2020

FinChat: Corpus and evaluation setup for Finnish chat conversations on everyday topics

Creating open-domain chatbots requires large amounts of conversational d...
research
01/12/2022

Human Evaluation of Conversations is an Open Problem: comparing the sensitivity of various methods for evaluating dialogue agents

At the heart of improving conversational AI is the open problem of how t...
research
10/05/2020

Spot The Bot: A Robust and Efficient Framework for the Evaluation of Conversational Dialogue Systems

The lack of time-efficient and reliable evaluation methods hamper the de...
research
12/28/2017

On the Challenges of Detecting Rude Conversational Behaviour

In this study, we aim to identify moments of rudeness between two indivi...

Please sign up or login with your details

Forgot password? Click here to reset