Transformer-based Localization from Embodied Dialog with Large-scale Pre-training

10/10/2022
by   Meera Hahn, et al.
0

We address the challenging task of Localization via Embodied Dialog (LED). Given a dialog from two agents, an Observer navigating through an unknown environment and a Locator who is attempting to identify the Observer's location, the goal is to predict the Observer's final location in a map. We develop a novel LED-Bert architecture and present an effective pretraining strategy. We show that a graph-based scene representation is more effective than the top-down 2D maps used in prior works. Our approach outperforms previous baselines.

READ FULL TEXT

page 1

page 7

research
11/16/2020

Where Are You? Localization from Embodied Dialog

We present Where Are You? (WAY), a dataset of  6k dialogs in which two h...
research
03/05/2020

EmpTransfo: A Multi-head Transformer Architecture for Creating Empathetic Dialog Systems

Understanding emotions and responding accordingly is one of the biggest ...
research
12/04/2021

Representation Learning for Conversational Data using Discourse Mutual Information Maximization

Although many pretrained models exist for text or images, there have bee...
research
06/02/2019

Pretraining Methods for Dialog Context Representation Learning

This paper examines various unsupervised pretraining objectives for lear...
research
12/05/2019

Large-scale Pretraining for Visual Dialog: A Simple State-of-the-Art Baseline

Prior work in visual dialog has focused on training deep neural models o...
research
10/28/2019

A Simple but Effective BERT Model for Dialog State Tracking on Resource-Limited Systems

In a task-oriented dialog system, the goal of dialog state tracking (DST...
research
09/30/2019

A Dynamic Strategy Coach for Effective Negotiation

Negotiation is a complex activity involving strategic reasoning, persuas...

Please sign up or login with your details

Forgot password? Click here to reset