Policy Transfer across Visual and Dynamics Domain Gaps via Iterative Grounding

07/01/2021
by   Grace Zhang, et al.
0

The ability to transfer a policy from one environment to another is a promising avenue for efficient robot learning in realistic settings where task supervision is not available. This can allow us to take advantage of environments well suited for training, such as simulators or laboratories, to learn a policy for a real robot in a home or office. To succeed, such policy transfer must overcome both the visual domain gap (e.g. different illumination or background) and the dynamics domain gap (e.g. different robot calibration or modelling error) between source and target environments. However, prior policy transfer approaches either cannot handle a large domain gap or can only address one type of domain gap at a time. In this paper, we propose a novel policy transfer method with iterative "environment grounding", IDAPT, that alternates between (1) directly minimizing both visual and dynamics domain gaps by grounding the source environment in the target environment domains, and (2) training a policy on the grounded source environment. This iterative training progressively aligns the domains between the two environments and adapts the policy to the target environment. Once trained, the policy can be directly executed on the target environment. The empirical results on locomotion and robotic manipulation tasks demonstrate that our approach can effectively transfer a policy across visual and dynamics domain gaps with minimal supervision and interaction with the target environment. Videos and code are available at https://clvrai.com/idapt .

READ FULL TEXT

page 1

page 6

page 13

page 14

page 16

research
10/12/2018

Policy Transfer with Strategy Optimization

Computer simulation provides an automatic and safe way for training robo...
research
05/14/2020

How to Close Sim-Real Gap? Transfer with Segmentation!

One fundamental difficulty in robotic learning is the sim-real gap probl...
research
08/04/2020

Stochastic Grounded Action Transformation for Robot Learning in Simulation

Robot control policies learned in simulation do not often transfer well ...
research
06/15/2021

Learning Stable Classifiers by Transferring Unstable Features

We study transfer learning in the presence of spurious correlations. We ...
research
12/11/2020

Protective Policy Transfer

Being able to transfer existing skills to new situations is a key capabi...
research
03/05/2018

Kinematic Morphing Networks for Manipulation Skill Transfer

The transfer of a robot skill between different geometric environments i...
research
06/20/2020

Counterfactually Guided Policy Transfer in Clinical Settings

Reliably transferring treatment policies learned in one clinical environ...

Please sign up or login with your details

Forgot password? Click here to reset