Multi-Stage Based Feature Fusion of Multi-Modal Data for Human Activity Recognition

11/08/2022
by   Hyeongju Choi, et al.
0

To properly assist humans in their needs, human activity recognition (HAR) systems need the ability to fuse information from multiple modalities. Our hypothesis is that multimodal sensors, visual and non-visual tend to provide complementary information, addressing the limitations of other modalities. In this work, we propose a multi-modal framework that learns to effectively combine features from RGB Video and IMU sensors, and show its robustness for MMAct and UTD-MHAD datasets. Our model is trained in two-stage, where in the first stage, each input encoder learns to effectively extract features, and in the second stage, learns to combine these individual features. We show significant improvements of 22 setup on UTD-MHAD dataset, and 20 experimentation, we show the robustness of our model on zero shot setting, and limited annotated data setting. We further compare with state-of-the-art methods that use more input modalities and show that our method outperforms significantly on the more difficult MMact dataset, and performs comparably in UTD-MHAD dataset.

READ FULL TEXT

page 2

page 4

research
08/13/2019

MEx: Multi-modal Exercises Dataset for Human Activity Recognition

MEx: Multi-modal Exercises Dataset is a multi-sensor, multi-modal datase...
research
12/04/2019

Template co-updating in multi-modal human activity recognition systems

Multi-modal systems are quite common in the context of human activity re...
research
05/05/2023

Distilled Mid-Fusion Transformer Networks for Multi-Modal Human Activity Recognition

Human Activity Recognition is an important task in many human-computer c...
research
10/14/2022

MMTSA: Multimodal Temporal Segment Attention Network for Efficient Human Activity Recognition

Multimodal sensors (e.g., visual, non-visual, and wearable) provide comp...
research
01/09/2019

Adaptive Feature Processing for Robust Human Activity Recognition on a Novel Multi-Modal Dataset

Human Activity Recognition (HAR) is a key building block of many emergin...
research
10/29/2019

Model enhancement and personalization using weakly supervised learning for multi-modal mobile sensing

Always-on sensing of mobile device user's contextual information is crit...
research
04/26/2022

Multi stain graph fusion for multimodal integration in pathology

In pathology, tissue samples are assessed using multiple staining techni...

Please sign up or login with your details

Forgot password? Click here to reset