A discussion on the validation tests employed to compare human action recognition methods using the MSR Action3D dataset

07/28/2014
by   José Ramón Padilla-López, et al.
0

This paper aims to determine which is the best human action recognition method based on features extracted from RGB-D devices, such as the Microsoft Kinect. A review of all the papers that make reference to MSR Action3D, the most used dataset that includes depth information acquired from a RGB-D device, has been performed. We found that the validation method used by each work differs from the others. So, a direct comparison among works cannot be made. However, almost all the works present their results comparing them without taking into account this issue. Therefore, we present different rankings according to the methodology used for the validation in orden to clarify the existing confusion.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 3

page 4

06/24/2019

A Comparative Review of Recent Kinect-based Action Recognition Algorithms

Video-based human action recognition is currently one of the most active...
11/27/2019

Literature Review of Action Recognition in the Wild

The literature review presented below on Action Recognition in the wild ...
01/21/2016

RGB-D-based Action Recognition Datasets: A Survey

Human action recognition from RGB-D (Red, Green, Blue and Depth) data ha...
01/27/2019

Spatio-temporal Action Recognition: A Survey

The task of action recognition or action detection involves analyzing vi...
12/08/2019

View-invariant Deep Architecture for Human Action Recognition using late fusion

Human action Recognition for unknown views is a challenging task. We pro...
10/16/2020

Pose And Joint-Aware Action Recognition

Most human action recognition systems typically consider static appearan...
09/28/2018

Overview of PicTropes, a film trope dataset

From the database DBTropes.org, we have created a dataset of films and t...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.