Community Regularization of Visually-Grounded Dialog

08/10/2018
by   Akshat Agarwal, et al.
0

The task of conducting visually grounded dialog involves learning goal-oriented cooperative dialog between autonomous agents who exchange information about a scene through several rounds of questions and answers in natural language. We posit that requiring artificial agents to adhere to the rules of human language, while also requiring them to maximize information exchange through dialog is an ill-posed problem. We observe that humans do not stray from a common language because they are social creatures who live in communities, and have to communicate with many people everyday, so it is far easier to stick to a common language even at the cost of some efficiency loss. Using this as inspiration, we propose and evaluate a multi-agent community-based dialog framework where each agent interacts with, and learns from, multiple agents, and show that this community-enforced regularization results in more relevant and coherent dialog (as judged by human evaluators) without sacrificing task performance (as judged by quantitative metrics).

READ FULL TEXT
research
08/10/2018

Mind Your Language: Learning Visually Grounded Dialog in a Multi-Agent Setting

The task of visually grounded dialog involves learning goal-oriented coo...
research
07/24/2020

Dialog without Dialog Data: Learning Visual Dialog Agents from VQA Data

Can we develop visually grounded dialog agents that can efficiently adap...
research
03/20/2017

Learning Cooperative Visual Dialog Agents with Deep Reinforcement Learning

We introduce the first goal-driven training for visual question answerin...
research
05/25/2022

The Dialog Must Go On: Improving Visual Dialog via Generative Self-Training

Visual dialog (VisDial) is a task of answering a sequence of questions g...
research
02/02/2022

Knowledge Engineering in the Long Game of Artificial Intelligence: The Case of Speech Acts

This paper describes principles and practices of knowledge engineering t...
research
06/26/2021

Saying the Unseen: Video Descriptions via Dialog Agents

Current vision and language tasks usually take complete visual data (e.g...
research
03/30/2023

DERA: Enhancing Large Language Model Completions with Dialog-Enabled Resolving Agents

Large language models (LLMs) have emerged as valuable tools for many nat...

Please sign up or login with your details

Forgot password? Click here to reset