Transferring ConvNet Features from Passive to Active Robot Self-Localization: The Use of Ego-Centric and World-Centric Views

04/22/2022
by   Kanya Kurauchi, et al.
0

The training of a next-best-view (NBV) planner for visual place recognition (VPR) is a fundamentally important task in autonomous robot navigation, for which a typical approach is the use of visual experiences that are collected in the target domain as training data. However, the collection of a wide variety of visual experiences in everyday navigation is costly and prohibitive for real-time robotic applications. We address this issue by employing a novel domain-invariant NBV planner. A standard VPR subsystem based on a convolutional neural network (CNN) is assumed to be available, and its domain-invariant state recognition ability is proposed to be transferred to train the domain-invariant NBV planner. Specifically, we divide the visual cues that are available from the CNN model into two types: the output layer cue (OLC) and intermediate layer cue (ILC). The OLC is available at the output layer of the CNN model and aims to estimate the state of the robot (e.g., the robot viewpoint) with respect to the world-centric view coordinate system. The ILC is available within the middle layers of the CNN model as a high-level description of the visual content (e.g., a saliency image) with respect to the ego-centric view. In our framework, the ILC and OLC are mapped to a state vector and subsequently used to train a multiview NBV planner via deep reinforcement learning. Experiments using the public NCLT dataset validate the effectiveness of the proposed method.

READ FULL TEXT

page 3

page 8

page 10

research
02/23/2021

Domain-invariant NBV Planner for Active Cross-domain Self-localization

Pole-like landmark has received increasing attention as a domain-invaria...
research
05/10/2023

Active Semantic Localization with Graph Neural Embedding

Semantic localization, i.e., robot self-localization with semantic image...
research
09/06/2021

Deep SIMBAD: Active Landmark-based Self-localization Using Ranking -based Scene Descriptor

Landmark-based robot self-localization has recently garnered interest as...
research
03/04/2020

Learning View and Target Invariant Visual Servoing for Navigation

The advances in deep reinforcement learning recently revived interest in...
research
03/12/2018

Omnidirectional CNN for Visual Place Recognition and Navigation

Visual place recognition is challenging, especially when only a few pla...
research
03/30/2021

Recognizing Actions in Videos from Unseen Viewpoints

Standard methods for video recognition use large CNNs designed to captur...
research
01/27/2022

Head and eye egocentric gesture recognition for human-robot interaction using eyewear cameras

Non-verbal communication plays a particularly important role in a wide r...

Please sign up or login with your details

Forgot password? Click here to reset