LazyDAgger: Reducing Context Switching in Interactive Imitation Learning

by   Ryan Hoque, et al.

Corrective interventions while a robot is learning to automate a task provide an intuitive method for a human supervisor to assist the robot and convey information about desired behavior. However, these interventions can impose significant burden on a human supervisor, as each intervention interrupts other work the human is doing, incurs latency with each context switch between supervisor and autonomous control, and requires time to perform. We present LazyDAgger, which extends the interactive imitation learning (IL) algorithm SafeDAgger to reduce context switches between supervisor and autonomous control. We find that LazyDAgger improves the performance and robustness of the learned policy during both learning and execution while limiting burden on the supervisor. Simulation experiments suggest that LazyDAgger can reduce context switches by an average of 60 while maintaining state-of-the-art policy performance. In physical fabric manipulation experiments with an ABB YuMi robot, LazyDAgger reduces context switches by 60 execution time.


page 1

page 4

page 5

page 6

page 9


ThriftyDAgger: Budget-Aware Novelty and Risk Gating for Interactive Imitation Learning

Effective robot learning often requires online human feedback and interv...

Human-in-the-Loop Imitation Learning using Remote Teleoperation

Imitation Learning is a promising paradigm for learning complex robot ma...

iCOIL: Scenario Aware Autonomous Parking Via Integrated Constrained Optimization and Imitation Learning

Autonomous parking (AP) is an emering technique to navigate an intellige...

Automated Task-Time Interventions to Improve Teamwork using Imitation Learning

Effective human-human and human-autonomy teamwork is critical but often ...

Disturbance Injection under Partial Automation: Robust Imitation Learning for Long-horizon Tasks

Partial Automation (PA) with intelligent support systems has been introd...

Learning Flight Control Systems from Human Demonstrations and Real-Time Uncertainty-Informed Interventions

This paper describes a methodology for learning flight control systems f...

IIFL: Implicit Interactive Fleet Learning from Heterogeneous Human Supervisors

Imitation learning has been applied to a range of robotic tasks, but can...

Please sign up or login with your details

Forgot password? Click here to reset