Indoor Future Person Localization from an Egocentric Wearable Camera

03/06/2021
by   Jianing Qiu, et al.
0

Accurate prediction of future person location and movement trajectory from an egocentric wearable camera can benefit a wide range of applications, such as assisting visually impaired people in navigation, and the development of mobility assistance for people with disability. In this work, a new egocentric dataset was constructed using a wearable camera, with 8,250 short clips of a targeted person either walking 1) toward, 2) away, or 3) across the camera wearer in indoor environments, or 4) staying still in the scene, and 13,817 person bounding boxes were manually labelled. Apart from the bounding boxes, the dataset also contains the estimated pose of the targeted person as well as the IMU signal of the wearable camera at each time point. An LSTM-based encoder-decoder framework was designed to predict the future location and movement trajectory of the targeted person in this egocentric setting. Extensive experiments have been conducted on the new dataset, and have shown that the proposed method is able to reliably and better predict future person location and trajectory in egocentric videos captured by the wearable camera compared to three baselines.

READ FULL TEXT

page 4

page 5

page 6

research
11/01/2021

Egocentric Human Trajectory Forecasting with a Wearable Camera and Multi-Modal Fusion

In this paper, we address the problem of forecasting the trajectory of a...
research
11/30/2017

Future Person Localization in First-Person Videos

We present a new task that predicts future locations of people observed ...
research
03/29/2018

Joint Person Segmentation and Identification in Synchronized First- and Third-person Videos

In a world in which cameras are becoming more and more pervasive, scenes...
research
04/17/2021

Wide-Baseline Multi-Camera Calibration using Person Re-Identification

We address the problem of estimating the 3D pose of a network of cameras...
research
12/01/2021

Pose2Room: Understanding 3D Scenes from Human Activities

With wearable IMU sensors, one can estimate human poses from wearable de...
research
08/12/2023

Fusion-GRU: A Deep Learning Model for Future Bounding Box Prediction of Traffic Agents in Risky Driving Videos

To ensure the safe and efficient navigation of autonomous vehicles and a...
research
11/10/2020

Simple means Faster: Real-Time Human Motion Forecasting in Monocular First Person Videos on CPU

We present a simple, fast, and light-weight RNN based framework for fore...

Please sign up or login with your details

Forgot password? Click here to reset