Learning Visual Servoing with Deep Features and Fitted Q-Iteration

03/31/2017
by   Alex X. Lee, et al.
0

Visual servoing involves choosing actions that move a robot in response to observations from a camera, in order to reach a goal configuration in the world. Standard visual servoing approaches typically rely on manually designed features and analytical dynamics models, which limits their generalization capability and often requires extensive application-specific feature and model engineering. In this work, we study how learned visual features, learned predictive dynamics models, and reinforcement learning can be combined to learn visual servoing mechanisms. We focus on target following, with the goal of designing algorithms that can learn a visual servo using low amounts of data of the target in question, to enable quick adaptation to new targets. Our approach is based on servoing the camera in the space of learned visual features, rather than image pixels or manually-designed keypoints. We demonstrate that standard deep features, in our case taken from a model trained for object classification, can be used together with a bilinear predictive model to learn an effective visual servo that is robust to visual variation, changes in viewing angle and appearance, and occlusions. A key component of our approach is to use a sample-efficient fitted Q-iteration algorithm to learn which features are best suited for the task at hand. We show that we can learn an effective visual servo on a complex synthetic car following benchmark using just 20 training trajectory samples for reinforcement learning. We demonstrate substantial improvement over a conventional approach based on image pixels or hand-designed keypoints, and we show an improvement in sample-efficiency of more than two orders of magnitude over standard model-free deep reinforcement learning algorithms. Videos are available at http://rll.berkeley.edu/visual_servoing .

READ FULL TEXT

page 4

page 7

page 8

page 16

research
12/30/2020

Model-Based Visual Planning with Self-Supervised Functional Distances

A generalist robot must be able to complete a variety of tasks in its en...
research
01/20/2022

DFBVS: Deep Feature-Based Visual Servo

Classical Visual Servoing (VS) rely on handcrafted visual features, whic...
research
10/07/2016

Deep Reinforcement Learning From Raw Pixels in Doom

Using current reinforcement learning methods, it has recently become pos...
research
10/10/2011

Closed-Loop Learning of Visual Control Policies

In this paper we present a general, flexible framework for learning mapp...
research
05/26/2019

Interactive Differentiable Simulation

Intelligent agents need a physical understanding of the world to predict...
research
09/29/2018

Robot eye-hand coordination learning by watching human demonstrations: a task function approximation approach

We present a robot eye-hand coordination learning method that can direct...
research
06/23/2016

Learning to Poke by Poking: Experiential Learning of Intuitive Physics

We investigate an experiential learning paradigm for acquiring an intern...

Please sign up or login with your details

Forgot password? Click here to reset