Verbal Focus-of-Attention System for Learning-from-Demonstration

07/17/2020
by   Naoki Wake, et al.
0

The Learning-from-Demonstration (LfD) framework aims to map human demonstrations to a robot to reduce programming effort. To this end, an LfD system encodes a human demonstration into a series of execution units for a robot, referred to as task models. Although previous research has proposed successful task-model encoders that analyze images and human body movements, the encoders have been designed in environments without noise. Therefore, there has been little discussion on how to guide a task-model encoder in a scene with spatio-temporal noises such as cluttered objects or unrelated human body movements. In human-to-human demonstrations, verbal instructions play a role in guiding an observer's visual attention. Inspired by the function of verbal instructions, we propose a verbal focus-of-attention (FoA) system (i.e., spatio-temporal filters) to guide a task-model encoder. For object manipulation, the encoder first recognizes a target-object name and its attributes from verbal instructions. The information serves as a where-to-look FoA filter to confine the areas where the target object existed in the demonstration. The encoder next detects the timings of grasp and release tasks that occur in the filtered area. The timings serve as a when-to-look FoA filter to confine the period when the demonstrator manipulated the object. Finally, the task-model encoder recognizes task models by employing the FoA filters. The contributions of this paper are: (1) to propose verbal FoA for LfD; (2) to design an algorithm to calculate FoA filters from verbal input; (3) to demonstrate the effectiveness of a verbal-driven FoA by testing an implemented LfD system in noisy environments.

READ FULL TEXT

page 1

page 7

research
12/21/2022

Interactive Learning-from-Observation through multimodal human demonstration

Learning-from-Observation (LfO) is a robot teaching framework for progra...
research
09/18/2023

DFL-TORO: A One-Shot Demonstration Framework for Learning Time-Optimal Robotic Manufacturing Tasks

This paper presents DFL-TORO, a novel Demonstration Framework for Learni...
research
05/16/2023

Interactive and Incremental Learning of Spatial Object Relations from Human Demonstrations

Humans use semantic concepts such as spatial relations between objects t...
research
11/28/2022

Proactive Robot Assistance via Spatio-Temporal Object Modeling

Proactive robot assistance enables a robot to anticipate and provide for...
research
10/10/2022

Using Both Demonstrations and Language Instructions to Efficiently Learn Robotic Tasks

Demonstrations and natural language instructions are two common ways to ...
research
07/12/2023

BiRP: Learning Robot Generalized Bimanual Coordination using Relative Parameterization Method on Human Demonstration

Human bimanual manipulation can perform more complex tasks than a simple...
research
05/12/2022

Robot Cooking with Stir-fry: Bimanual Non-prehensile Manipulation of Semi-fluid Objects

This letter describes an approach to achieve well-known Chinese cooking ...

Please sign up or login with your details

Forgot password? Click here to reset