Vision-Guided Robot Hearing

11/06/2013
by   Xavier Alameda-Pineda, et al.
0

Natural human-robot interaction in complex and unpredictable environments is one of the main research lines in robotics. In typical real-world scenarios, humans are at some distance from the robot and the acquired signals are strongly impaired by noise, reverberations and other interfering sources. In this context, the detection and localisation of speakers plays a key role since it is the pillar on which several tasks (e.g.: speech recognition and speaker tracking) rely. We address the problem of how to detect and localize people that are both seen and heard by a humanoid robot. We introduce a hybrid deterministic/probabilistic model. Indeed, the deterministic component allows us to map the visual information into the auditory space. By means of the probabilistic component, the visual features guide the grouping of the auditory features in order to form AV objects. The proposed model and the associated algorithm are implemented in real-time (17 FPS) using a stereoscopic camera pair and two microphones embedded into the head of the humanoid robot NAO. We performed experiments on (i) synthetic data, (ii) a publicly available data set and (iii) data acquired using the robot. The results we obtained validate the approach and encourage us to further investigate how vision can help robot hearing.

READ FULL TEXT

page 3

page 5

page 14

page 15

page 17

page 20

page 22

research
06/15/2013

iCub World: Friendly Robots Help Building Good Vision Data-Sets

In this paper we present and start analyzing the iCub World data-set, an...
research
12/30/2017

Multichannel Robot Speech Recognition Database: MChRSR

In real human robot interaction (HRI) scenarios, speech recognition repr...
research
06/17/2019

Weighted delay-and-sum beamforming guided by visual tracking for human-robot interaction

This paper describes the integration of weighted delay-and-sum beamformi...
research
07/22/2023

Estimating speaker direction on a humanoid robot with binaural acoustic signals

To achieve human-like behaviour during speech interactions, it is necess...
research
11/12/2020

Self-supervised reinforcement learning for speaker localisation with the iCub humanoid robot

In the future robots will interact more and more with humans and will ha...
research
06/06/2022

People Tracking in Panoramic Video for Guiding Robots

A guiding robot aims to effectively bring people to and from specific pl...

Please sign up or login with your details

Forgot password? Click here to reset