Learning under Misspecified Objective Spaces

10/11/2018
by   Andreea Bobu, et al.
0

Learning robot objective functions from human input has become increasingly important, but state-of-the-art techniques assume that the human's desired objective lies within the robot's hypothesis space. When this is not true, even methods that keep track of uncertainty over the objective fail because they reason about which hypothesis might be correct, and not whether any of the hypotheses are correct. We focus specifically on learning from physical human corrections during the robot's task execution, where not having a rich enough hypothesis space leads to the robot updating its objective in ways that the person did not actually intend. We observe that such corrections appear irrelevant to the robot, because they are not the best way of achieving any of the candidate objectives. Instead of naively trusting and learning from every human interaction, we propose robots learn conservatively by reasoning in real time about how relevant the human's correction is for the robot's hypothesis space. We test our inference method in an experiment with human interaction data, and demonstrate that this alleviates unintended learning in an in-person user study with a 7DoF robot manipulator.

READ FULL TEXT
research
02/03/2020

Quantifying Hypothesis Space Misspecification in Learning from Human-Robot Demonstrations and Physical Corrections

Human input has enabled autonomous systems to improve their capabilities...
research
07/06/2021

Physical Interaction as Communication: Learning Robot Objectives Online from Human Corrections

When a robot performs a task next to a human, physical interaction is in...
research
11/09/2020

Joint Estimation of Expertise and Reward Preferences From Human Demonstrations

When a robot learns from human examples, most approaches assume that the...
research
03/31/2021

Learning Human Objectives from Sequences of Physical Corrections

When personal, assistive, and interactive robots make mistakes, humans n...
research
03/09/2019

Literal or Pedagogic Human? Analyzing Human Model Misspecification in Objective Learning

It is incredibly easy for a system designer to misspecify the objective ...
research
10/12/2021

Auditing Robot Learning for Safety and Compliance during Deployment

Robots of the future are going to exhibit increasingly human-like and su...
research
01/13/2021

Top Program Construction and Reduction for polynomial time Meta-Interpretive Learning

Meta-Interpretive Learners, like most ILP systems, learn by searching fo...

Please sign up or login with your details

Forgot password? Click here to reset