Log In Sign Up

Tracking Ensemble Performance on Touch-Screens with Gesture Classification and Transition Matrices

by   Charles Martin, et al.

We present and evaluate a novel interface for tracking ensemble performances on touch-screens. The system uses a Random Forest classifier to extract touch-screen gestures and transition matrix statistics. It analyses the resulting gesture-state sequences across an ensemble of performers. A series of specially designed iPad apps respond to this real-time analysis of free-form gestural performances with calculated modifications to their musical interfaces. We describe our system and evaluate it through cross-validation and profiling as well as concert experience.


page 1

page 4


A Dynamic Modelling Framework for Human Hand Gesture Task Recognition

Gesture recognition and hand motion tracking are important tasks in adva...

Comparing Hand Gestures and the Gamepad Interfaces for Locomotion in Virtual Environments

Hand gesture is a new and promising interface for locomotion in virtual ...

Real-Time Online Skeleton Extraction and Gesture Recognition on Pepper

We present a multi-stage pipeline for simple gesture recognition. The no...

Progression Modelling for Online and Early Gesture Detection

Online and Early detection of gestures is crucial for building touchless...

Fast Gesture Recognition with Multiple Stream Discrete HMMs on 3D Skeletons

HMMs are widely used in action and gesture recognition due to their impl...

Classification of Viral Pneumonia X-ray Images with the Aucmedi Framework

In this work we use the AUCMEDI-Framework to train a deep neural network...

Cross-artform performance using networked interfaces: Last Man to Die's Vital LMTD

In 2009 the cross artform group, Last Man to Die, presented a series of ...