Transporters with Visual Foresight for Solving Unseen Rearrangement Tasks

02/22/2022
by   Hongtao Wu, et al.
0

Rearrangement tasks have been identified as a crucial challenge for intelligent robotic manipulation, but few methods allow for precise construction of unseen structures. We propose a visual foresight model for pick-and-place manipulation which is able to learn efficiently. In addition, we develop a multi-modal action proposal module which builds on Goal-Conditioned Transporter Networks, a state-of-the-art imitation learning method. Our method, Transporters with Visual Foresight (TVF), enables task planning from image data and is able to achieve multi-task learning and zero-shot generalization to unseen tasks with only a handful of expert demonstrations. TVF is able to improve the performance of a state-of-the-art imitation learning method on both training and unseen tasks in simulation and real robot experiments. In particular, the average success rate on unseen tasks improves from 55.0 77.9 when given only tens of expert demonstrations. More details can be found on our project website: https://chirikjianlab.github.io/tvf/

READ FULL TEXT

page 1

page 2

page 3

page 5

page 7

research
02/04/2022

BC-Z: Zero-Shot Task Generalization with Robotic Imitation Learning

In this paper, we study the problem of enabling a vision-based robotic m...
research
10/26/2021

Towards More Generalizable One-shot Visual Imitation Learning

A general-purpose robot should be able to master a wide range of tasks a...
research
04/23/2018

Zero-Shot Visual Imitation

The current dominant paradigm for imitation learning relies on strong su...
research
03/05/2023

Seq2Seq Imitation Learning for Tactile Feedback-based Manipulation

Robot control for tactile feedback-based manipulation can be difficult d...
research
08/05/2020

Generalization Guarantees for Multi-Modal Imitation Learning

Control policies from imitation learning can often fail to generalize to...
research
01/18/2023

NeRF in the Palm of Your Hand: Corrective Augmentation for Robotics via Novel-View Synthesis

Expert demonstrations are a rich source of supervision for training visu...
research
05/30/2023

Language-Conditioned Imitation Learning with Base Skill Priors under Unstructured Data

The growing interest in language-conditioned robot manipulation aims to ...

Please sign up or login with your details

Forgot password? Click here to reset