Learning Category-Level Manipulation Tasks from Point Clouds with Dynamic Graph CNNs

09/13/2022
by   Junchi Liang, et al.
0

This paper presents a new technique for learning category-level manipulation from raw RGB-D videos of task demonstrations, with no manual labels or annotations. Category-level learning aims to acquire skills that can be generalized to new objects, with geometries and textures that are different from the ones of the objects used in the demonstrations. We address this problem by first viewing both grasping and manipulation as special cases of tool use, where a tool object is moved to a sequence of key-poses defined in a frame of reference of a target object. Tool and target objects, along with their key-poses, are predicted using a dynamic graph convolutional neural network that takes as input an automatically segmented depth and color image of the entire scene. Empirical results on object manipulation tasks with a real robotic arm show that the proposed network can efficiently learn from real visual demonstrations to perform the tasks on novel objects within the same category, and outperforms alternative approaches.

READ FULL TEXT

page 1

page 6

page 8

research
03/08/2022

Learning Sensorimotor Primitives of Sequential Manipulation Tasks from Visual Demonstrations

This work aims to learn how to perform complex robot manipulation tasks ...
research
07/30/2021

ManiSkill: Learning-from-Demonstrations Benchmark for Generalizable Manipulation Skills

Learning generalizable manipulation skills is central for robots to achi...
research
03/07/2018

Adapting Everyday Manipulation Skills to Varied Scenarios

We address the problem of executing tool-using manipulation skills in sc...
research
05/07/2022

Category-Independent Articulated Object Tracking with Factor Graphs

Robots deployed in human-centric environments may need to manipulate a d...
research
02/07/2023

Local Neural Descriptor Fields: Locally Conditioned Object Representations for Manipulation

A robot operating in a household environment will see a wide range of un...
research
06/16/2022

Equivariant Descriptor Fields: SE(3)-Equivariant Energy-Based Models for End-to-End Visual Robotic Manipulation Learning

End-to-end learning for visual robotic manipulation is known to suffer f...
research
07/31/2020

Telemanipulation with Chopsticks: Analyzing Human Factors in User Demonstrations

Chopsticks constitute a simple yet versatile tool that humans have used ...

Please sign up or login with your details

Forgot password? Click here to reset