Self-Adapting Recurrent Models for Object Pushing from Learning in Simulation

07/27/2020
by   Lin Cong, et al.
0

Planar pushing remains a challenging research topic, where building the dynamic model of the interaction is the core issue. Even an accurate analytical dynamic model is inherently unstable because physics parameters such as inertia and friction can only be approximated. Data-driven models usually rely on large amounts of training data, but data collection is time consuming when working with real robots. In this paper, we collect all training data in a physics simulator and build an LSTM-based model to fit the pushing dynamics. Domain Randomization is applied to capture the pushing trajectories of a generalized class of objects. When executed on the real robot, the trained recursive model adapts to the tracked object's real dynamics within a few steps. We propose the algorithm Recurrent Model Predictive Path Integral (RMPPI) as a variation of the original MPPI approach, employing state-dependent recurrent models. As a comparison, we also train a Deep Deterministic Policy Gradient (DDPG) network as a model-free baseline, which is also used as the action generator in the data collection phase. During policy training, Hindsight Experience Replay is used to improve exploration efficiency. Pushing experiments on our UR5 platform demonstrate the model's adaptability and the effectiveness of the proposed framework.

READ FULL TEXT

page 1

page 2

page 6

research
01/20/2021

Active Model Learning using Informative Trajectories for Improved Closed-Loop Control on Real Robots

Model-based controllers on real robots require accurate knowledge of the...
research
06/16/2020

Data Driven Control with Learned Dynamics: Model-Based versus Model-Free Approach

This paper compares two different types of data-driven control methods, ...
research
02/27/2018

Friction Variability in Auto-collected Dataset of Planar Pushing Experiments and Anisotropic Friction

Friction plays a key role in manipulating objects. Most of what we do wi...
research
03/03/2019

Asynchronous Episodic Deep Deterministic Policy Gradient: Towards Continuous Control in Computationally Complex Environments

Deep Deterministic Policy Gradient (DDPG) has been proved to be a succes...
research
02/28/2022

A Recurrent Differentiable Engine for Modeling Tensegrity Robots Trainable with Low-Frequency Data

Tensegrity robots, composed of rigid rods and flexible cables, are diffi...
research
07/18/2023

Sampling-based Model Predictive Control Leveraging Parallelizable Physics Simulations

We present a method for sampling-based model predictive control that mak...
research
11/06/2022

Leveraging Haptic Feedback to Improve Data Quality and Quantity for Deep Imitation Learning Models

Learning from demonstration (LfD) is a proven technique to teach robots ...

Please sign up or login with your details

Forgot password? Click here to reset