Developing Motion Code Embedding for Action Recognition in Videos

12/10/2020
by   Maxat Alibayev, et al.
4

In this work, we propose a motion embedding strategy known as motion codes, which is a vectorized representation of motions based on a manipulation's salient mechanical attributes. These motion codes provide a robust motion representation, and they are obtained using a hierarchy of features called the motion taxonomy. We developed and trained a deep neural network model that combines visual and semantic features to identify the features found in our motion taxonomy to embed or annotate videos with motion codes. To demonstrate the potential of motion codes as features for machine learning tasks, we integrated the extracted features from the motion embedding model into the current state-of-the-art action recognition model. The obtained model achieved higher accuracy than the baseline model for the verb classification task on egocentric videos from the EPIC-KITCHENS dataset.

READ FULL TEXT
research
07/31/2020

Estimating Motion Codes from Demonstration Videos

A motion taxonomy can encode manipulations as a binary-encoded represent...
research
05/09/2019

A Taxonomy and Dataset for 360° Videos

In this paper, we propose a taxonomy for 360 videos that categorizes vid...
research
07/28/2016

SEMBED: Semantic Embedding of Egocentric Action Videos

We present SEMBED, an approach for embedding an egocentric object intera...
research
07/13/2020

A Motion Taxonomy for Manipulation Embedding

To represent motions from a mechanical point of view, this paper explore...
research
09/12/2023

Self-supervised Extraction of Human Motion Structures via Frame-wise Discrete Features

The present paper proposes an encoder-decoder model for extracting the s...
research
10/01/2019

Manipulation Motion Taxonomy and Coding for Robots

This paper introduces a taxonomy of manipulations as seen especially in ...
research
04/30/2022

RADNet: A Deep Neural Network Model for Robust Perception in Moving Autonomous Systems

Interactive autonomous applications require robustness of the perception...

Please sign up or login with your details

Forgot password? Click here to reset