One-Shot Imitation Learning

03/21/2017
by   Yan Duan, et al.
0

Imitation learning has been commonly applied to solve different tasks in isolation. This usually requires either careful feature engineering, or a significant number of samples. This is far from what we desire: ideally, robots should be able to learn from very few demonstrations of any given task, and instantly generalize to new situations of the same task, without requiring task-specific engineering. In this paper, we propose a meta-learning framework for achieving such capability, which we call one-shot imitation learning. Specifically, we consider the setting where there is a very large set of tasks, and each task has many instantiations. For example, a task could be to stack all blocks on a table into a single tower, another task could be to place all blocks on a table into two-block towers, etc. In each case, different instances of the task would consist of different sets of blocks with different initial states. At training time, our algorithm is presented with pairs of demonstrations for a subset of all tasks. A neural net is trained that takes as input one demonstration and the current state (which initially is the initial state of the other demonstration of the pair), and outputs an action with the goal that the resulting sequence of states and actions matches as closely as possible with the second demonstration. At test time, a demonstration of a single instance of a new task is presented, and the neural net is expected to perform well on new instances of this new task. The use of soft attention allows the model to generalize to conditions and tasks unseen in the training data. We anticipate that by training this model on a much greater variety of tasks and settings, we will obtain a general system that can turn any demonstrations into robust policies that can accomplish an overwhelming variety of tasks. Videos available at https://bit.ly/nips2017-oneshot .

READ FULL TEXT

page 18

page 19

page 26

page 27

research
10/17/2022

Robust Imitation of a Few Demonstrations with a Backwards Model

Behavior cloning of expert demonstrations can speed up learning optimal ...
research
12/04/2018

Compositional Imitation Learning: Explaining and executing one task at a time

We introduce a framework for Compositional Imitation Learning and Execut...
research
08/30/2023

RoboTAP: Tracking Arbitrary Points for Few-Shot Visual Imitation

For robots to be useful outside labs and specialized factories we need a...
research
09/15/2019

State Representation Learning from Demonstration

In a context where several policies can be observed as black boxes on di...
research
11/23/2019

Meta Adaptation using Importance Weighted Demonstrations

Imitation learning has gained immense popularity because of its high sam...
research
08/15/2023

Leveraging Symmetries in Pick and Place

Robotic pick and place tasks are symmetric under translations and rotati...
research
02/09/2023

One-shot Visual Imitation via Attributed Waypoints and Demonstration Augmentation

In this paper, we analyze the behavior of existing techniques and design...

Please sign up or login with your details

Forgot password? Click here to reset