Let's Dance: Learning From Online Dance Videos

01/23/2018
by   Daniel Castro, et al.
0

In recent years, deep neural network approaches have naturally extended to the video domain, in their simplest case by aggregating per-frame classifications as a baseline for action recognition. A majority of the work in this area extends from the imaging domain, leading to visual-feature heavy approaches on temporal data. To address this issue we introduce "Let's Dance", a 1000 video dataset (and growing) comprised of 10 visually overlapping dance categories that require motion for their classification. We stress the important of human motion as a key distinguisher in our work given that, as we show in this work, visual information is not sufficient to classify motion-heavy categories. We compare our datasets' performance using imaging techniques with UCF-101 and demonstrate this inherent difficulty. We present a comparison of numerous state-of-the-art techniques on our dataset using three different representations (video, optical flow and multi-person pose data) in order to analyze these approaches. We discuss the motion parameterization of each of them and their value in learning to categorize online dance videos. Lastly, we release this dataset (and its three representations) for the research community to use.

READ FULL TEXT

page 1

page 3

page 5

page 6

page 7

research
07/20/2020

MotionSqueeze: Neural Motion Feature Learning for Video Understanding

Motion plays a crucial role in understanding videos and most state-of-th...
research
09/10/2018

Using phase instead of optical flow for action recognition

Currently, the most common motion representation for action recognition ...
research
10/12/2019

Generating Human Action Videos by Coupling 3D Game Engines and Probabilistic Graphical Models

Deep video action recognition models have been highly successful in rece...
research
10/30/2018

Random Temporal Skipping for Multirate Video Analysis

Current state-of-the-art approaches to video understanding adopt tempora...
research
07/06/2012

Analysis of Multi-Scale Fractal Dimension to Classify Human Motion

In recent years there has been considerable interest in human action rec...
research
09/06/2016

Making a Case for Learning Motion Representations with Phase

This work advocates Eulerian motion representation learning over the cur...
research
12/01/2018

From Third Person to First Person: Dataset and Baselines for Synthesis and Retrieval

First-person (egocentric) and third person (exocentric) videos are drast...

Please sign up or login with your details

Forgot password? Click here to reset