Generating Virtual On-body Accelerometer Data from Virtual Textual Descriptions for Human Activity Recognition

05/04/2023
by   Zikang Leng, et al.
3

The development of robust, generalized models in human activity recognition (HAR) has been hindered by the scarcity of large-scale, labeled data sets. Recent work has shown that virtual IMU data extracted from videos using computer vision techniques can lead to substantial performance improvements when training HAR models combined with small portions of real IMU data. Inspired by recent advances in motion synthesis from textual descriptions and connecting Large Language Models (LLMs) to various AI models, we introduce an automated pipeline that first uses ChatGPT to generate diverse textual descriptions of activities. These textual descriptions are then used to generate 3D human motion sequences via a motion synthesis model, T2M-GPT, and later converted to streams of virtual IMU data. We benchmarked our approach on three HAR datasets (RealWorld, PAMAP2, and USC-HAD) and demonstrate that the use of virtual IMU training data generated using our new approach leads to significantly improved HAR model performance compared to only using real IMU data. Our approach contributes to the growing field of cross-modality transfer methods and illustrate how HAR models can be improved through the generation of virtual training data that do not require any manual effort.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/02/2022

Fine-grained Human Activity Recognition Using Virtual On-body Acceleration Data

Previous work has demonstrated that virtual accelerometry data, extracte...
research
07/21/2020

Creating a Large-scale Synthetic Dataset for Human Activity Recognition

Creating and labelling datasets of videos for use in training Human Acti...
research
06/16/2023

Vision-Language Models can Identify Distracted Driver Behavior from Naturalistic Videos

Recognizing the activities, causing distraction, in real-world driving s...
research
11/23/2020

Yet it moves: Learning from Generic Motions to Generate IMU data from YouTube videos

Human activity recognition (HAR) using wearable sensors has benefited mu...
research
09/07/2023

A Function Interpretation Benchmark for Evaluating Interpretability Methods

Labeling neural network submodules with human-legible descriptions is us...
research
04/29/2020

Effective Human Activity Recognition Based on Small Datasets

Most recent work on vision-based human activity recognition (HAR) focuse...
research
08/31/2023

Expanding Frozen Vision-Language Models without Retraining: Towards Improved Robot Perception

Vision-language models (VLMs) have shown powerful capabilities in visual...

Please sign up or login with your details

Forgot password? Click here to reset