Task-Driven Out-of-Distribution Detection with Statistical Guarantees for Robot Learning

06/25/2021
by   Alec Farid, et al.
8

Our goal is to perform out-of-distribution (OOD) detection, i.e., to detect when a robot is operating in environments that are drawn from a different distribution than the environments used to train the robot. We leverage Probably Approximately Correct (PAC)-Bayes theory in order to train a policy with a guaranteed bound on performance on the training distribution. Our key idea for OOD detection then relies on the following intuition: violation of the performance bound on test environments provides evidence that the robot is operating OOD. We formalize this via statistical techniques based on p-values and concentration inequalities. The resulting approach (i) provides guaranteed confidence bounds on OOD detection, and (ii) is task-driven and sensitive only to changes that impact the robot's performance. We demonstrate our approach on a simulated example of grasping objects with unfamiliar poses or shapes. We also present both simulation and hardware experiments for a drone performing vision-based obstacle avoidance in unfamiliar environments (including wind disturbances and different obstacle densities). Our examples demonstrate that we can perform task-driven OOD detection within just a handful of trials. Comparisons with baselines also demonstrate the advantages of our approach in terms of providing statistical guarantees and being insensitive to task-irrelevant distribution shifts.

READ FULL TEXT
research
02/28/2020

Probably Approximately Correct Vision-Based Planning using Motion Primitives

This paper presents a deep reinforcement learning approach for synthesiz...
research
11/16/2021

Stronger Generalization Guarantees for Robot Learning by Combining Generative Models and Real-World Data

We are motivated by the problem of learning policies for robotic systems...
research
02/11/2022

Failure Prediction with Statistical Guarantees for Vision-Based Robot Control

We are motivated by the problem of performing failure prediction for saf...
research
11/16/2021

Learning Provably Robust Motion Planners Using Funnel Libraries

This paper presents an approach for learning motion planners that are ac...
research
07/13/2021

Distributionally Robust Policy Learning via Adversarial Environment Generation

Our goal is to train control policies that generalize well to unseen env...
research
06/11/2018

PAC-Bayes Control: Synthesizing Controllers that Provably Generalize to Novel Environments

Our goal is to synthesize controllers for robots that provably generaliz...
research
09/20/2018

Task-Driven Estimation and Control via Information Bottlenecks

Our goal is to develop a principled and general algorithmic framework fo...

Please sign up or login with your details

Forgot password? Click here to reset