Leveraging Multiple Environments for Learning and Decision Making: a Dismantling Use Case

by   Alejandro Suárez-Hernández, et al.

Learning is usually performed by observing real robot executions. Physics-based simulators are a good alternative for providing highly valuable information while avoiding costly and potentially destructive robot executions. We present a novel approach for learning the probabilities of symbolic robot action outcomes. This is done leveraging different environments, such as physics-based simulators, in execution time. To this end, we propose MENID (Multiple Environment Noise Indeterministic Deictic) rules, a novel representation able to cope with the inherent uncertainties present in robotic tasks. MENID rules explicitly represent each possible outcomes of an action, keep memory of the source of the experience, and maintain the probability of success of each outcome. We also introduce an algorithm to distribute actions among environments, based on previous experiences and expected gain. Before using physics-based simulations, we propose a methodology for evaluating different simulation settings and determining the least time-consuming model that could be used while still producing coherent results. We demonstrate the validity of the approach in a dismantling use case, using a simulation with reduced quality as simulated system, and a simulation with full resolution where we add noise to the trajectories and some physical parameters as a representation of the real system.



There are no comments yet.


page 2

page 6

page 7


From Video Game to Real Robot: The Transfer between Action Spaces

Training agents with reinforcement learning based techniques requires th...

robosuite: A Modular Simulation Framework and Benchmark for Robot Learning

robosuite is a simulation framework for robot learning powered by the Mu...

Interactive simulation for easy decision-making in fluid dynamics

A conventional study of fluid simulation involves different stages inclu...

RFUniverse: A Physics-based Action-centric Interactive Environment for Everyday Household Tasks

Household environments are important testbeds for embodied AI research. ...

Provenance, Anonymisation and Data Environments: a Unifying Construction

The Anonymisation Decision-making Framework (ADF) operationalizes the ri...

Learning the Noise of Failure: Intelligent System Tests for Robots

Roboticists usually test new control software in simulation environments...

Leveraging Reinforcement Learning Techniques for Effective Policy Adoption and Validation

Rewards and punishments in different forms are pervasive and present in ...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.