Neural Task Graphs: Generalizing to Unseen Tasks from a Single Video Demonstration

07/10/2018
by   De-An Huang, et al.
2

Our goal is for a robot to execute a previously unseen task based on a single video demonstration of the task. The success of our approach relies on the principle of transferring knowledge from seen tasks to unseen ones with similar semantics. More importantly, we hypothesize that to successfully execute a complex task from a single video demonstration, it is necessary to explicitly incorporate compositionality to the model. To test our hypothesis, we propose Neural Task Graph (NTG) Networks, which use task graph as the intermediate representation to modularize the representations of both the video demonstration and the derived policy. We show this formulation achieves strong inter-task generalization on two complex tasks: Block Stacking in BulletPhysics and Object Collection in AI2-THOR. We further show that the same principle is applicable to real-world videos. We show that NTG can improve data efficiency of few-shot activity understanding in the Breakfast Dataset.

READ FULL TEXT

page 2

page 11

page 12

page 13

page 14

research
10/04/2017

Neural Task Programming: Learning to Generalize Across Hierarchical Tasks

In this work, we propose a novel robot learning framework called Neural ...
research
09/29/2019

Learning from Observations Using a Single Video Demonstration and Human Feedback

In this paper, we present a method for learning from video demonstration...
research
07/17/2023

Video-Mined Task Graphs for Keystep Recognition in Instructional Videos

Procedural activity understanding requires perceiving human actions in t...
research
03/09/2022

One-Shot Learning from a Demonstration with Hierarchical Latent Language

Humans have the capability, aided by the expressive compositionality of ...
research
10/13/2022

Augmentation for Learning From Demonstration with Environmental Constraints

We introduce a Learning from Demonstration (LfD) approach for contact-ri...
research
10/17/2018

One-Shot Observation Learning

Observation learning is the process of learning a task by observing an e...
research
12/21/2016

Temporal Tessellation: A Unified Approach for Video Analysis

We present a general approach to video understanding, inspired by semant...

Please sign up or login with your details

Forgot password? Click here to reset