WEAR: A Multimodal Dataset for Wearable and Egocentric Video Activity Recognition

04/11/2023
by   Marius Bock, et al.
0

Though research has shown the complementarity of camera- and inertial-based data, datasets which offer both modalities remain scarce. In this paper we introduce WEAR, a multimodal benchmark dataset for both vision- and wearable-based Human Activity Recognition (HAR). The dataset comprises data from 18 participants performing a total of 18 different workout activities with untrimmed inertial (acceleration) and camera (egocentric video) data recorded at 10 different outside locations. WEAR features a diverse set of activities which are low in inter-class similarity and, unlike previous egocentric datasets, not defined by human-object-interactions nor originate from inherently distinct activity categories. Provided benchmark results reveal that single-modality architectures have different strengths and weaknesses in their prediction performance. Further, in light of the recent success of transformer-based video action detection models, we demonstrate their versatility by applying them in a plain fashion using vision, inertial and combined (vision + inertial) features as input. Results show that vision transformers are not only able to produce competitive results using only inertial data, but also can function as an architecture to fuse both modalities by means of simple concatenation, with the multimodal approach being able to produce the highest average mAP, precision and close-to-best F1-scores. Up until now, vision-based transformers have neither been explored in inertial nor in multimodal human activity recognition, making our approach the first to do so. The dataset and code to reproduce experiments is publicly available via: mariusbock.github.io/wear

READ FULL TEXT
research
05/22/2023

Hang-Time HAR: A Benchmark Dataset for Basketball Activity Recognition using Wrist-worn Inertial Sensors

We present a benchmark dataset for evaluating physical human activity re...
research
09/12/2022

BON: An extended public domain dataset for human activity recognition

Body-worn first-person vision (FPV) camera enables to extract a rich sou...
research
03/02/2023

MuscleMap: Towards Video-based Activated Muscle Group Estimation

In this paper, we tackle the new task of video-based Activated Muscle Gr...
research
02/08/2023

InMyFace: Inertial and Mechanomyography-Based Sensor Fusion for Wearable Facial Activity Recognition

Recognizing facial activity is a well-understood (but non-trivial) compu...
research
07/14/2022

Inertial Hallucinations – When Wearable Inertial Devices Start Seeing Things

We propose a novel approach to multimodal sensor fusion for Ambient Assi...
research
07/06/2020

ARC-Net: Activity Recognition Through Capsules

Human Activity Recognition (HAR) is a challenging problem that needs adv...
research
01/17/2022

Homogenization of Existing Inertial-Based Datasets to Support Human Activity Recognition

Several techniques have been proposed to address the problem of recogniz...

Please sign up or login with your details

Forgot password? Click here to reset