Learning Articulated Motions From Visual Demonstration

by   Sudeep Pillai, et al.

Many functional elements of human homes and workplaces consist of rigid components which are connected through one or more sliding or rotating linkages. Examples include doors and drawers of cabinets and appliances; laptops; and swivel office chairs. A robotic mobile manipulator would benefit from the ability to acquire kinematic models of such objects from observation. This paper describes a method by which a robot can acquire an object model by capturing depth imagery of the object as a human moves it through its range of motion. We envision that in future, a machine newly introduced to an environment could be shown by its human user the articulated objects particular to that environment, inferring from these "visual demonstrations" enough information to actuate each object independently of the user. Our method employs sparse (markerless) feature tracking, motion segmentation, component pose estimation, and articulation learning; it does not require prior object models. Using the method, a robot can observe an object being exercised, infer a kinematic model incorporating rigid, prismatic and revolute joints, then use the model to predict the object's motion from a novel vantage point. We evaluate the method's performance, and compare it to that of a previously published technique, for a variety of household objects.



There are no comments yet.


page 1

page 2

page 5

page 6


Inferring geometric constraints in human demonstrations

This paper presents an approach for inferring geometric constraints in h...

Learning Descriptor of Constrained Task from Demonstration

Constrained objects, such as doors and drawers are often complex and sha...

A Versatile Framework for Robust and Adaptive Door Operation with a Mobile Manipulator Robot

The ability to deal with articulated objects is very important for robot...

Learning to Infer Kinematic Hierarchies for Novel Object Instances

Manipulating an articulated object requires perceiving itskinematic hier...

Learning Periodic Tasks from Human Demonstrations

We develop a method for learning periodic tasks from visual demonstratio...

Online Tool and Task learning via Human Robot Interaction

This work describes the development of a robotic system that acquires kn...

A Method for Constraint Inference Using Pose and Wrench Measurements

Many physical tasks such as pulling out a drawer or wiping a table can b...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.