Teleoperator Imitation with Continuous-time Safety

05/23/2019
by   Bachir El Khadir, et al.
0

Learning to effectively imitate human teleoperators, with generalization to unseen and dynamic environments, is a promising path to greater autonomy enabling robots to steadily acquire complex skills from supervision. We propose a new motion learning technique rooted in contraction theory and sum-of-squares programming for estimating a control law in the form of a polynomial vector field from a given set of demonstrations. Notably, this vector field is provably optimal for the problem of minimizing imitation loss while providing continuous-time guarantees on the induced imitation behavior. Our method generalizes to new initial and goal poses of the robot and can adapt in real-time to dynamic obstacles during execution, with convergence to teleoperator behavior within a well-defined safety tube. We present an application of our framework for pick-and-place tasks in the presence of moving obstacles on a 7-DOF KUKA IIWA arm. The method compares favorably to other learning-from-demonstration approaches on benchmark handwriting imitation tasks.

READ FULL TEXT

page 1

page 2

page 6

research
02/25/2020

Scalable Multi-Task Imitation Learning with Autonomous Improvement

While robot learning has demonstrated promising results for enabling rob...
research
06/22/2020

PICO: Primitive Imitation for COntrol

In this work, we explore a novel framework for control of complex system...
research
01/29/2022

Robust Imitation Learning from Corrupted Demonstrations

We consider offline Imitation Learning from corrupted demonstrations whe...
research
07/08/2021

Imitation by Predicting Observations

Imitation learning enables agents to reuse and adapt the hard-won expert...
research
06/15/2021

Causal Navigation by Continuous-time Neural Networks

Imitation learning enables high-fidelity, vision-based learning of polic...

Please sign up or login with your details

Forgot password? Click here to reset